Learn how to evaluate your Agno Agents and Teams for accuracy using LLM-as-a-judge methodology with input/output pairs.
AccuracyEval
will run the Agent with the input, then use a different model (o4-mini
) to score the Agent’s response according to the guidelines provided.
AccuracyEval
with tools.
Create a virtual environment
Terminal
and create a python virtual environment.Install libraries
Run Basic Accuracy Example
Test Accuracy with Tools
Test with Given Answer
Test Async Accuracy
Test Team Accuracy
Test Number Comparison
Run the Evals Demo
View the Evals Demo