Review test results
This section guides you through the business workflow for reviewing test results. This workflow is designed for business users who need to review evaluation results, understand failures, and determine the appropriate actions to take.
Starting reviews
There are two main ways to review test results:
From an evaluation run
From an assigned task
From an evaluation run
When reviewing a failure directly from a test execution (not from a task), follow these steps:
Review a fail after a test execution - After a test execution, review the failure details
Determine the appropriate action - Based on your review, decide which of the following scenarios applies:
graph LR
A[Review Failure] --> B{Agent Answer<br/>Correct?}
B -->|No| C[<a href="task_management.html" target="_self">Open Task<br/>Assign to Developer<br/>or KB Manager</a>]
B -->|Yes| F{Rewrite Now?}
B -->|Don't Know| E[<a href="task_management.html" target="_self">Put in Draft<br/>Open Task<br/>Assign to Domain Expert</a>]
F -->|Yes| G{Can Answer<br/>Questions?}
F -->|No| H[<a href="task_management.html" target="_self">Draft Test Case<br/>Create Task<br/>Assign to PO</a>]
G -->|Yes| I[<a href="modify_test_cases.html" target="_self">Rewrite Test<br/>Retest<br/>Save</a>]
G -->|No| J{Has Value?}
J -->|No| K[Remove Test]
J -->|Yes| H
Tip
To review evaluation runs, you first need to run an evaluation. For information on running evaluations, see Run and review evaluations. For information on viewing evaluation results, see Run, review, schedule and compare evaluation runs.
If the agent is incorrect, the test is well written
If the agent is incorrect and the test is correctly identifying the issue:
Open a task and assign the agent developer or the KB manager
Navigate to the “Distribute tasks” workflow Distribute tasks to organize your review work
Create a task with a clear description of what needs to be fixed
If the agent is correct, the test should be rewritten
If the agent is correct and the test was too strict, you need to rewrite the test. You have the following options:
Option 1: You want to do it later
Draft the test case - Mark the test case as draft to prevent it from being used in evaluations
Open a task where you can track that this test case needs to be modified
Assign the product owner to the task
Navigate to the “Distribute tasks” workflow Distribute tasks to organize your review work
Option 2: You are able to answer at least one of these questions:
Is there any minimum information the agent must not omit (e.g., a number, a fact)?
Is there any block of information the agent must not go beyond (a page of a website, a section of a document)?
Is there any information you do not want to appear in the agent’s answer?
If you can answer at least one of these questions:
Go to the linked test case in the dataset
Rewrite the test requirement:
If question 1 is true: Enable correctness check by putting the minimum info as reference
If question 2 is true: Enable groundedness check and put the block of info as context
If question 3 is true: Write a negative rule (“the agent should not…”) in a conformity check
Retest various times until the result is always PASS (regenerate a agent answer, and retest)
Save the changes
If the test case was in draft, undraft it
You can also set the task as closed (if applicable)
Option 3: The test does not have value
Remove it from the dataset
Tip
For detailed information about modifying test cases, see Modify the test cases.
If you don’t know, there needs to be a discussion
If you don’t know if the agent answers correctly or not and there needs to be a discussion:
Put in draft - Mark the test case as draft to prevent it from being used in evaluations
Open a task and assign the domain expert
Navigate to the “Distribute tasks” workflow Distribute tasks to organize your review work
Create a task with your questions and concerns, then assign it to the domain expert who can make this determination
From an assigned task
When reviewing a task that has been assigned to you, follow these steps:
Open the task - Open the task that has been assigned to you
Read the failure details - Review the description, result, and explanation for the failure
Determine the appropriate action - Based on your review, decide which of the following scenarios applies:
graph LR
B[Review Failure] --> C{Agent Answer<br/>Correct?}
C -->|No| D[<a href="task_management.html" target="_self">Assign to Developer</a>]
C -->|Yes| E[<a href="task_management.html" target="_self">Update Task Description<br/>Assign to Product Owner</a>]
C -->|Don't Know| F[<a href="task_management.html" target="_self">Update Task Description<br/>Assign to Expert or PO</a>]
Tip
For information on creating tasks, see Distribute tasks to organize your review work.
If the agent is incorrect, the test is well written
Assign the task to the developer who should correct the test
Navigate to the “Distribute tasks” workflow Distribute tasks to organize your review work
Reassign the task to the appropriate developer with a clear description of what needs to be fixed
If the agent is correct, the test should be rewritten
If the agent answers correctly in reality and the test was too strict:
Provide the reason why the agent answer is ok, in the description of the task
Answer at least one of these questions to help guide the test rewrite:
Is there any minimum information the agent must not omit (e.g., a number, a fact)?
Is there any block of information the agent must not go beyond (a page of a website, a section of a document)?
Is there any information you do not want to appear in the agent’s answer?
Assign the product owner so that he or she can rewrite the test based on your input
Navigate to the “Distribute tasks” workflow Distribute tasks to organize your review work
Update the task description with your answer and reassign it to the product owner
If you don’t know if the agent answers correctly or not. There needs to be a discussion
If you don’t know if the agent answers correctly or not and there needs to be a discussion:
Provide the reason why you don’t know and why it needs to be discussed
Assign the right person with the knowledge or re-assign the product owner
Navigate to the “Distribute tasks” workflow Distribute tasks to organize your review work
Update the task with your questions and concerns, then reassign it to the appropriate person
Interpreting test results
Check pass/fail
When reviewing a test case, the first thing to check is whether the test case passed or failed. By opening the test case, you can see the metrics along with the failure category and tags on the right side of the screen.
PASS:
The test case met all the evaluation criteria (checks)
All checks that were enabled on the test case passed
The agent’s response was acceptable according to the validation rules
FAIL:
The test case did not meet one or more evaluation criteria
At least one check that was enabled on the test case failed
To understand why a test case failed, you need to review the specific checks that were applied.
Tip
For detailed information about checks and how they work, see Understand metrics, failure categories and tags. For information on enabling/disabling checks, see the “Enable/Disable checks” section in Modify the test cases.
Check failure reason
To understand why a test passed or failed, you need to review the explanation for each check and understand the failure categories.
Read the explanation for each check
Each check provides an explanation of why it passed or failed. This explanation helps you understand:
What the check was evaluating
What criteria were applied
Why the test case passed or failed
What specific aspects of the agent’s response caused the result
Tip
For more information about checks and how to enable/disable them, see the “Enable/Disable checks” section in Modify the test cases. For comprehensive information about all check types, see Understand metrics, failure categories and tags.
Check failure category
When a test fails, it is categorized based on the type of failure. Understanding these categories helps you:
Identify patterns in failures
Prioritize which issues to address first
Assign tasks to the right team members
Common failure categories:
Hallucination - The agent generated information not present in the context
Omission - The agent failed to include required information
Conformity violation - The agent did not follow business rules or constraints
Groundedness issue - The agent’s answer contains information not grounded in the provided context
Metadata mismatch - The agent’s metadata does not match expected values
String matching failure - Required keywords or phrases are missing
Tip
You can change the categories used for classification but before doing so, we recommend you to read about the best practices for modifying test cases in Modify the test cases.
Review the flow of the conversation
Understanding the conversation flow helps you assess whether the test case structure is appropriate and whether the agent’s response makes sense in context.
When reviewing the conversation flow, consider:
Whether the conversation structure makes sense
Whether the user messages are clear and unambiguous
Whether the conversation history provides necessary context
Whether the test case accurately represents the scenario you want to test
Conversation structure
A conversation, or test case, is composed of a sequence of messages between the user and the assistant, alternating between each role. When designing your test cases, you can provide conversation history by adding multiple turns (multi-turn), but the conversation should always end with a user message. The agent’s next assistant completion will be generated and evaluated at test case time.
Simple conversation
In the simplest scenario, a conversation consists of a single message from the user. For example:
User: Hello, which language is your open-source library written in?
Multi-turn conversation
To test multi-turn capabilities or provide more context, you can add several alternating messages. For instance:
User: Hello, I wanted to have more information about your open-source library.
Assistant: Hello! I’m happy to help you learn more about our library. What would you like to know?
User: Which language is it written in?
You can add as many turns as needed, but always ensure the conversation ends with a user message, since the assistant’s reply will be evaluated at runtime.
Conversation Answer Examples
You can also provide an “answer example” for each test case. This answer example is not used during evaluation runs, but helps when annotating the dataset or validating your evaluation criteria. For example, you might want to:
Import answer examples together with conversations by providing a demo_output field in your dataset.
Generate the agent’s answer by replacing the assistant message directly in the interface.
Write your own answer example to check specific behaviors or validation rules.
If you do not provide an answer example, the Hub will automatically use the assistant reply generated during the first evaluation run as the default example.
Tip
For more detailed information about creating and managing conversations, see Review test results.
Conversation metadata
The conversation metadata provides additional information about the agent’s response, which a developer decided to pass along with the answer, such as:
Tool calls that were made
System flags or status indicators
Additional context or structured data
Any other information the agent includes in its response
Reviewing metadata helps you understand:
What actions the agent took during the conversation
Whether the agent followed expected workflows
Whether system-level requirements were met
Whether the response structure matches expectations
For more information about metadata checks and other check types, see Understand metrics, failure categories and tags.
Best practices
Review thoroughly - Take time to understand all aspects of the test result before making a decision
Document your findings - Add comments to tasks to help others understand your review
Use appropriate actions - Close tasks when results are correct, assign modification work when changes are needed
Collaborate effectively - Work with product owners and other team members to ensure test cases are accurate
Maintain quality - Only close tasks when you’re confident the test results are correct
Next steps
Now that you understand how to review test results, you can:
Modify test cases - Learn how to refine test cases and checks Modify the test cases
Distribute tasks - Create and manage tasks to organize review work Distribute tasks to organize your review work