Agenta is an open-source LLMOps platform for building reliable AI apps. Manage prompts, run evaluations, and debug traces. We help developers and domain experts collaborate to ship LLM applications faster and with confidence.
I really like Agenta and chose it as my #1 tool for prompt engineering when I researched different tools. I needed a tool to teach my students at the University of Calgary how to do systematic prompt engineering studies, and this one was the best one for a non-technical audience to access all the professional tools for such studies in one place. I'm planning to get our PhD students into it now for prompt engineering studies that can turn into full-fledged research papers.
Been juggling prompts in git and evals in notebooks. Agenta looks like the boring tooling I actually want. Open source is nice. Trace debug view = clutch. If it plays nice with PostHog/LangChain, I’m in. Saving this for next sprint.
I could see how our tool could benefit from Agenta.ai’s agent-orchestration workflow to streamline complex automation tasks across our platform, and I’m definitely going to take a closer look at their launch to explore how we can make this work together.
seldomly can see an opensource project for LLMops like Agenta! Great launch and congrats team!
@mahmoudmabrouk Love the open-source approach! 🚀
How do you handle prompt versioning when multiple teams are collaborating? Can you roll back to previous versions easily?
Huge congrats! Love seeing more open-source tooling that actually helps teams ship with confidence.
@mabrouk Love seeing more momentum in the LLMOps space, especially with an open-source approach. Most teams trying to ship AI features hit the same wall: lots of prompts, zero visibility, and no reliable way to evaluate or debug what’s actually happening under the hood.
A platform that unifies prompts, evals, and trace debugging feels like a real unlock for both devs and domain experts who don’t want to depend on guesswork.
Curious: what’s been the biggest challenge so far, capturing consistent traces, defining evaluation metrics, or helping teams collaborate around prompt changes?
This looks pretty handy finally a place for AI teams to manage prompts, test them, and debug without chaos. Open-source makes it even better, feels like a tool teams can actually build on and trust.
As a PM, have been trying several tools for evals, super excited to try this one!!
Oh wow, this is really amazing. Collaborating with the team on prompts and debugging with evaluations is a really cool idea. It seems like AI tools are really evolving :) Also, I see APIs, and that makes it even more exciting. Would love to try that out.
I recently evaluated Agenta vs Langfuse for Prompt Management and tracing. I went with Langfuse this time but all the best for this project. Open Source FTW.
Something that would really set you apart, that no one else seems to have, would be approval workflows for Prompt management. Managing prompts in the UI is great but in a remotely business-y environment I can't let one person have the ability to push new prompts without checks and balances. We'll probably have to manage this with source control (e.g. Github) and write some script to push prompts up to Langfuse once they gain approval.
Hi there!
Agenta is a workspace where AI teams collaborate effectively to build reliable AI applications.
Whether you’re building interactive chat apps, single-prompt workflows, or more agentic systems, Agenta keeps everything in one place instead of having prompts, experiments, and evaluations scattered across different tools.
We’d really appreciate your feedback or ideas, and thanks for taking the time to check it out at cloud.agenta.ai (free forever) or contact us at agenta-hq.slack.com for a demo !
Nice to see a tool that lets both devs and non-tech team members collaborate. Best wishes to the team. One thing I am curious how Agent handles versioning for prompts and evaluations?
Hi everyone! 👋 We built Agenta to have a way for AI teams to collaborate on prompts. We offer a complete workflow for building reliable AI apps, form prompt engineering, to evaluation and observability. We'd love to hear your thoughts, feedback or ideas — thanks for checking us out! 🙌
Hi Product Hunt 👋
I'm Mahmoud, co-founder of Agenta. The team and I are excited to launch Agenta today.
What is Agenta?
Agenta is an open-source platform that helps AI teams ship reliable LLM applications.
The Problem
Building a demo is easy. Building a reliable app is hard.
Small prompt changes improve one case but break another
Subject matter experts and engineers can't collaborate easily (prompts end up scattered across code and spreadsheets)
Teams don't know if their prompts are working in production
How Agenta Solves This
Playground for the whole team. Everyone can experiment with prompts and models, not just engineers.
Deploy without code changes. Anyone can push a working prompt instantly.
Test before you ship. Create test cases and validate prompts against them (no more vibe-based prompting).
Monitor in production. Track mistakes, user feedback and costs after deployment.
Who's Using Agenta
Hundreds of teams use Agenta Cloud (generous free tier) or self-host it. They run more experiments, ship AI features faster, and collaborate in one place.
Great product! Can I integrate prompts from it to my app via API/SDK?
Can I use variables in the prompt?