# Red Teaming Test your AI system against adversarial attacks using automated red teaming techniques. Simulate real-world attack scenarios to identify vulnerabilities and weaknesses in your system prompt defense. AI Attack Simulation Red Teaming gives you automated adversarial testing with multiple attack techniques. After the testing is complete, a report is generated with the success rates and the interactions that occurred during testing. To help you get started, there are pre-configured red team prompt sets provided. See [Red Team Prompt Sets](/docs/products/console/attack_simulation_red_teaming_prompt_sets) for more information. ## What you get - **Automated adversarial testing** across multiple attack techniques and tactics. - **Comprehensive attack reports** with success rates, failure reasons, and detailed interactions. - **Objective-based testing** to verify your system prompt's robustness against specific threats. ## When to use Red Teaming - Validating security of an AI system. - Testing effectiveness of security controls. - Comparing different prompt versions. - Compliance and security auditing.