Agentuity Documentation

Evals

Docs

Evaluations are automated quality checks that run after your agent responds. They don't slow down your response; they execute in the background and results appear in the Agentuity console. Two types: binary (pass/fail) and score (0-1). Use preset evals like answer-completeness or create custom evals with your own logic. Evals help you catch quality issues before users do and track performance over time.

Explainer
OpenAI/gpt-5-nano
Evaluator
Groq/gpt-oss-120b
Prompt
Explain what AI is and how it works in a few brief sentences
Reference Code
Loading...
Ready
Output will appear here...