November 10 - 14
Launch Week

Monday, 10 November
See how your prompts perform across all metrics at a glance
Compare prompt versions side by side to spot regressions fast
Debug with complete traces to understand every output
Customize LLM-as-a-judge evaluators with any schema you need
Tuesday, 12 November
Live view of the reliability of your system in production
Gain confidence that your outputs meet your quality standards
Find edge cases and add them to your test cases to improve your AI system
Clear insight into how prompt changes behave in production
Wednesday, 13 November
Create or fetch test sets programatically
Write custom evaluators or use buiilt-in evaluators
Evaluate end to end or specific steps
View results in the dashboard
Thursday, 14 November
All functional features now open source (MIT license)
Includes evaluation, prompt management, and observability
Development back in the public repo
Need help?
What is Agenta?
Agenta is an open-source LLMOps platform that includes all the tool to adopt the best practices for building reliable LLM powered applications. It includes tools for prompt management, prompt engineering, LLM evaluation, and observability.





