Skip to content
GitHubDiscord

Overview

Test your AI agent for safety and security vulnerabilities with automated red teaming attacks.

The vulnerability scan helps you identify weaknesses in your AI agent by testing it against common attack patterns. This includes:

  • Prompt injection attempts
  • Harmful content generation
  • Data extraction attacks
  • Other OWASP GenAI Top 10 risks

How it works: The scan runs dozens of specialized red teaming probes that adapt to your agent’s capabilities and use case. Each probe tests for specific vulnerabilities and provides detailed results.

What you get:

  • A security grade (A-D) based on detected vulnerabilities
  • Detailed breakdown by attack category and severity
  • Conversation logs showing exactly how attacks were performed
  • Actionable insights to improve your agent’s defenses

Example of vulnerability scan results

graph LR
    A([<a href="/hub/ui/scan/launch-scan">Launch Scan</a>]) --> C([<a href="/hub/ui/scan/review-scan-results">Review Vulnerabilities</a>])
    C --> D{Take Action}
    D -->|Convert to Test| E[Send to Dataset]
    D -->|Create Task| F[<a href="/hub/ui/annotate/task-management">Distribute Task</a>]
    E --> H[<a href="/hub/ui/annotate/modify-test-cases">Review Test Cases</a>]
    F --> H
    H --> A

The scan tests for these common AI security risks: