Cekura is an end-to-end QA for Voice & Chat AI Agents. Cekura helps Conversational AI companies with pre-production testing and simulation as well as monitoring of production calls to ensure quality and reliability at every stage of development
Cekura lets you simulate, evaluate and monitor your Voice & Chat AI agents automatically.
Why did we build Cekura? 💡
Cekura was born out of our own frustration building voice agents for healthcare, where every change required hours of manual QA, yet critical failures still made it to production. We built the platform we wished existed: one that simulates conversations at scale, generates edge-case scenarios, and monitors real-world agent calls for failures.
Team Background 👥
Shashij has published a paper on AI systems testing from his research at ETH Zurich and Google. Tarush has developed simulations for ultra-low latency trading, and I have led product and growth teams before, including a conversational AI company. All of us met at IIT Bombay and have been friends for the last 8 years.
Problem🚨: Making Conversational AI agents reliable is hard. Manually calling/chatting with your agents or listening through thousands of conversations is slow, error-prone and does not provide the required coverage.
Our Solution: At Cekura, we work closely with you at each step of the agent-building journey and help you improve and scale your agents 10 times faster
Key Features:
Testing:
Scenario Generation: Create varied test cases from agent descriptions automatically for comprehensive coverage.
Evaluation Metrics: Track custom and AI-generated metrics. Check for instruction following, tool calls, and conversational metrics (Interruptions, Latency, etc).
Prompt Recommendation: Get actionable insights to improve each of the metrics.
Custom Personas: Emulate diverse user types with varied accents, background noise, and conversational styles.
Production Call & Chat Simulation: Simulate production calls to ensure all the fixes have been incorporated.
Instruction Following: Identify instances where agents fail to follow instructions.
Drop-off Tracking: Analyzes when and why users abandon calls, highlighting areas of improvement.
Custom Metrics: Define unique metrics for personalized call analysis.
Alerting: Proactively notifies users of critical issues like latency spikes or missed functions.
Major Updates Since Last Product Hunt Launch:
Added Chat AI Testing and Observability
Automated Expected Outcome along with each generated scenario
Simulation of Production conversations
'Instruction Following' and 'Hallucination' metric to automatically flag deviations from Agent description and Knowledge base respectively
Who is this for?
Anyone building Conversational AI agents. If you want to make your voice & chat AI agents reliable, book my calendar here 🗓️ or reach out to [email protected]📧.
If you'd like to engage in a fun roleplay, you can talk with our agent here: You will act as a customer support representative and our agent will call you for a refund, order status, and product recommendation. After the call, we will give you an evaluation.
Please note: In reality, we generate hundreds of simulations automatically and provide detailed analytics on your AI agent's performance as demonstrated in the demo video.
Hi folks! Great product and congrats on the launch! Thanks for sharing the demo and the role-play agent
I have a question from a business perspective: How important do you think is staying relevant on social media platforms for a B2B startup like Cekura? What are some things you have done that generated a lot of traction?
An unrelated question: what is an 80-20 feature of your product? Something that you added after feedback with not a lot of effort, but generated tremendous value for your clients?
Congratulations on the launch!
What kind of improvements have your customers seen after using Cekura AI?
Congratulation on the lauch.Best in the market and impressive.
Hi guys, congratulations for this incredible work. I have some questions -
How does pricing work—are plans usage‑based or seat‑based?
Do you support multichannel testing—e.g., SMS or chat—in addition to voice?
Congrats on the launch, go Cekura!!
Exactly what the next generation of apps need
Congratulations, my brother @shashij_gupta & TEAM on launching your product! That’s an incredible achievement. I’ve been curious - can Cekura detect and score tool calls (function calls) inside a conversation? Would love to learn more about its capabilities!.
Congratulations on the launch!
What are some of the metrics used, especially for the quality of Conversational AI agents?
Which voice/telephony platforms and protocols does Cekura plug into out of the box (Twilio, SIP, WebRTC, etc.)?
Do you support multiple languages and accents, including European markets?
Hi! I'm just curious to understand, if I’m already using Vapi or Retell, or ElevenLabs, where does Cekura add extra value?
Great one team, How customizable is the testing? May I upload my own scenarios, personas, and KPIs?
Congrats on the second launch Sid. I was wondering, what sets Cekura apart from other eval tools like Galileo or Braintrust? :)
What’s the missing piece that you think is present in current model evaluations that Cekura is solving/ will solve?
Great product, even better team. I did a thorough bakeoff of the options out there to improve dev velocity and stability for a voice agent, and Cekura had the most intuitive platform. But what really drove it home was their true customer obsession and top-notch service. They’ve been deep in the weeds with us and real thought partners in our voice agent journey.
What you guys do is crucial for the Voice AI industry !! Thank you for the effort and all the best for the launch and the following 🦾
We use Cekura for all our production applications. Highly recommend it for any enterprise or large agency building many and/or complex Voice AI applications.