August 6, 2025
Hamming is transforming AI voice agent testing with automated testing infrastructure that's 100x more efficient than manual testing.
Backed by NEA, Y Combinator, Pioneer Fund, and Mischief, Hamming enables teams to create thousands of concurrent phone calls to AI voice agents using realistic voice characters, helping companies catch regressions instantly.
As a company building critical infrastructure for fast-growing AI companies, maintaining high code quality is fundamental to their mission.
The Hamming engineering team were no strangers to software testing, especially given that they themselves build testing tools for top AI companies.
Sumanyu Sharma, Co-Founder & CEO of Hamming, recognized that their own codebase needed stronger testing practices. Any bugs in their core evaluation system could directly impact their customers' ability to test and deploy voice agents reliably.
The team faced the challenge of any high-growth company: they needed to ship fast while also backfilling test coverage across their codebase.
Writing unit tests manually for existing code risked consuming engineering time that would be better spent on customer-facing features. Moreover, identifying which files needed the most testing attention was becoming a chore as their repo grew.
"We needed a way to increase our test coverage without slowing down our dev cycles," explained Sumanyu. "The challenge beyond writing the tests itself was also knowing where to focus our efforts for maximum impact."
CoverBot has been ideal for our use case of getting test coverage up. We went from 2,500 tests on our core evals functionality to over 7,000+ tests in a month. CoverBot played a big part in achieving these gains.
Hamming decided to implement Tusk’s CoverBot, an AI-powered workflow product designed specifically for backfilling unit tests.
CoverBot acted as their AI teammate that would create PRs twice a day to cover existing code that had minimal tests. Hamming’s engineers could then offload the work of backfilling unit tests for certain directories to the bot.
Traditionally, backfilling would mean manually identifying under-tested code by running a coverage script. CoverBot automated that end-to-end process by analyzing their repo to identify the least-tested files before creating test suites with full codebase context.
CoverBot proved to be what Hamming needed during their dedicated sprint to improve unit test coverage. The key differentiators:
In three months, Hamming enjoyed these results:
Using CoverBot allowed their engineering team to focus on feature development while still getting the perks of robust testing. What would have taken months of consistent effort to address was shortened to a matter of days.
Get PRs automatically on a daily or weekly basis with high-quality tests that raise coverage. No coverage analysis or prompting required. To find out more about Tusk, book a demo with us.
For automated AI voice agent testing that's 100x faster than manual testing, visit Hamming to learn how their platform can help you ensure reliability in your voice agents at scale.