Compare evaluations
Comparing evaluations is a crucial part of maintaining and improving your LLM agents over time. By comparing results across different versions, datasets, or time periods, you can:
Detect regressions: Identify when agent performance has degraded
Track improvements: Measure the impact of changes and optimizations
Maintain quality standards: Ensure consistent performance across deployments
Make data-driven decisions: Use metrics to guide development priorities
In this section, we will walk you through how to compare evaluations in Giskard Hub.
How to compare evaluations
On the Evaluations page, select at least two evaluations to compare, then click the “Compare” button in the top right corner of the table. The page will display a comparison of the selected evaluations.

Understanding the comparison view
First, it shows the success rate - the percentage of conversations that the checks passed in each evaluation. It also displays the percentage of each specific check. Then it presents a table listing the conversations, which can be filtered by results, such as whether the conversations in agenth evaluations passed or failed the checks.
Conversation-level analysis
Clicking on a conversation will show a detailed comparison.

Within this comparison you can explore the performance of the agent on a specific conversation and metrics.
Tip
💡 How to use your test results to correct your AI agent?
During this process you might uncover patterns and issues that you can address in your agent.
For example, if you created a custom check to verify whether the agent starts with “I’m sorry,” it is useful to know how many conversations fail this requirement. If the failure rate is high, you can chose to adjust the evaluation, create more representative test cases or adjust your Agent deployment.
If you need more information on setting up efficient evaluations for your agent, check out the Evaluate tests and assign validation rules section.
Next steps
Now that you have compared evaluations, you can take action on the results.
Schedule evaluations - Schedule evaluations