As AI agents grow more complex, reasoning, using tools, and making decisions, traditional evals fall short. LangWatch Scenario simulates real-world interactions to test agent behavior. It’s like unit testing, but for AI agents.
We're excited to be launching LangWatch Scenario the first and only testing platform that allows you to test agents in simulated realities, with confidence and alongside domain expertise.
The problem that we’ve found is that teams are building increasingly complex agents, but testing them is still manual, time-consuming, and unreliable. You tweak a prompt, manually chat with your agent, hope it works better... and repeat. It's like shipping software without unit tests.
Our solution: Agent simulations that automatically test your AI agents across multiple scenarios. Think of it as a test suite for agents — catch regressions before they hit production, simulate edge cases alongside domain experts in a collaborative fashion, and ship with confidence.
What makes us different:
🧠 Agent simulations that act as unit tests for AI agents
🧪 Simulate multi-turn, edge-case scenarios
🧑💻 Code-first, no lock-in, framework-agnostic
👩⚕️ Built for domain experts and not just devs
🔍 Catch failures before users see them
✅ Trust your agent in production, not just evals
🏗️ Works with any agent framework (LangGraph, CrewAI, etc.)
LangWatch scenarios is our latest breakthrough that will allow teams to ship agents with confidence, not crossed fingers.
Love this shift, treating agents like software just makes sense. Do teams use it more pre- or post-deploy?
Congrats @manouk_dr & team, this is a huge step forward for reliable agent development. Feels like the missing test layer for AI. 👏
Congrats for the launch 🚀
Scenario testing seems like a game changer for the non-deterministic nature of AI. It's very cool to see testing and quality tools finally emerging for this new wave of agent-based systems.
I've known @r0bertp3rry since 2016 and he's always been an enthusiast of the ML field, I remember a chat bot demo of him while back when it wasn't even a thing everyone talked about. So, it's awesome to see him building in this space now.
Huge congrats to the team! 👏
evals and quick testing of agents is much needed. will give this product a go. congrats on the launch!
Hello everyone! 👋
I'm Rogerio, founder of LangWatch, been developing software for 15+ years, and my career really changed once I started dominating unit tests, TDD and so on, not only delivering mission critical software with zero bugs but also having a much more pleasant experience in doing so.
So I couldn't be more excited for the Agent Simulations solution we are bringing today to the world, it feels like finally the missing piece in delivering agents, bringing much stronger craftsmanship to agent development.
I'll be your technical guide here, ask me anything!