Hamming AI (YC S24)

Hamming AI (YC S24)

Automated testing for voice agents

5.0
β€’1 reviewβ€’

362 followers

Hamming tests your AI voice agents 100x faster than manual calls. Create Character.ai-style personas and scenarios. Run 100s of simultaneous phone calls to find bugs in your voice agents. Get detailed analytics on where to improve.
Hamming AI (YC S24) gallery image
Hamming AI (YC S24) gallery image
Hamming AI (YC S24) gallery image
Hamming AI (YC S24) gallery image
Hamming AI (YC S24) gallery image
Hamming AI (YC S24) gallery image
Hamming AI (YC S24) gallery image
Free Options
Launch Team / Built With
Unblocked AI Code Review
Unblocked AI Code Review
High-signal comments based on your team's context
Promoted

What do you think? …

Sumanyu Sharma
πŸ‘‹ Hi ya'll - Sumanyu and Marius here from Hamming AI. Hamming lets you automatically test your LLM voice agent. In our interactive demo, you play the role of the voice agent, and our agent will play the role of a difficult end user. We'll then score your performance on the call. πŸ•΅οΈ Try it here: https://app.hamming.ai/voice-demo (no signup needed). In practice, our agents call your agent! Marius and I previously ran growth and data teams at companies like Citizen, Tesla, and Anduril. We're excited to launch our automated voice testing feature to help you test your voice agents 100x faster than manual phone calls. πŸ“ž LLM voice agents currently require a LOT of iteration and tuning. For example, one of our customers is building an LLM drive-through voice agent for fast food chains. Their KPI is order accuracy. It's crucial for their system to gracefully handle dietary restrictions like allergies and customers who get distracted or otherwise change their minds mid-order. Mistakes in this context could lead to unhappy customers, potential health risks, and financial losses. πŸͺ„ Our solution involves four steps: (1) Create diverse but realistic user personas and scenarios covering the expected conversation space. We create these ourselves for each of our customers. (2) Have our agents call your agent when we test your agent's ability to handle things like background noise, long silences, or interruptions. Or have us test just the LLM / logic layer (function calls, etc.) via an API hook. (3) We score the outputs for each conversation using deterministic checks and LLM judges tailored to the specific problem domain (e.g., order accuracy, tone, friendliness). (4) Re-use the checks and judges above to score production traffic and use it to track quality metrics in production. (i.e., online evals) We created a Loom recording showing our customers' logged-in experience: Logged-in Video Walkthrough We think there will be more and more voice companies, and making the experimentation process easier is a problem we are excited about solving. πŸ“© If you're building voice agents and you're struggling to make them reliable, reach out at sumanyu@hamming.ai! ❀️ Shoutout to @rajiv_ayyangar and @gabe for helping us with the launch!
Rohan Chaubey
πŸ”Œ Plugged in
@sumanyu_sharma Congrats on the launch Sumanyu! :) This is the best first comment I have seen on PH outlining the product details, walkthrough, contact info., etc. Shared it with our community.
Sumanyu Sharma
@rohanrecommends You're the absolute best. Love your support!
Nikhil Pareek
Evals for a specific industry is a great idea. Llm as a judge is great, but comes with its own challenges, would be interesting to see how it performs for wide usecases. Also, persona generation automation based on usecases would also be great. Im sure thats in our roadmap πŸ˜€ Congratulations on the launch! πŸš€
Sumanyu Sharma
@nikhilpareek Absolutely. So far we've seen 95%+ alignment between LLM and human judgement. Yup, we're already doing persona generation based on use cases :)
Nikhil Pareek
@sumanyu_sharma thats awesome! Congratulations
Sumanyu Sharma
@nikhilpareek Try the demo here: https://app.hamming.ai/voice-demo We created these three to show the range but we can basically create any persona you can imagine :)
Tony Han
1000 parallel simulated calls to the AI voice agent is such a banger line and claim! As a product person, this is my #1 concern when I build AI product which is consistency. Sometimes, you just don't know why something breaks. Testing by hands and evaluating with eyes only go so far. It's about time for automated testing for LLM voice agent! Does this work with traditional chatbot usage? For example, we have one AI avatar talking to a human through web app? Congrats on the launch @sumanyu_sharma and team!
Sumanyu Sharma
@tonyhanded Yup it works for traditional conversational chat as well; voice is just a special case of the former!
Bailey Spell
Couldn't be more excited for these guys!! They have been working incredibly hard and have provided significant value to all of their current customers. They love them and anyone building in voice can get incredibly better results through Hamming!
Sumanyu Sharma
@baileyg2016 I appreciate your support! πŸ™
Illumi Killua
Hi @sumanyu_sharma congrats on the launch πŸ’₯
Sumanyu Sharma
@illumi_killua Thank you! 😊
Roop Reddy
@sumanyu_sharma Cheers for the launch!! Are the personas fixed or custom to my use case?
Sumanyu Sharma
@roopreddy Good question! We create bespoke personas for your use case. This way the simulators mimic how real customers interact with your systems!
Vidhi Nagpal
Impressive efficiency boost!
123
β€’β€’β€’
Next
Last