Skip to content

System Prompt Evaluation Summary

Provides summary details, like the number of requests and tokens. Also provides the original system prompt, the recommended compiled system prompt, and the security updates provided by the recommended system prompt.

Run a Red Team Evaluation against the original system prompt or the compiled system prompt.

Data Retention

System Prompt Evaluation data is retained for 90 days.

System Prompt Evaluation Details

Original System Prompt

The initial system prompt, provided by the user, when the evaluation was created.

Actions

Click the action button (three vertical dots) to view the Actions menu.

Create Red Team Evaluation using Original System Prompt

  1. To view to a system prompt evaluation summary, select Security Testing > System Prompt, then click the green arrow for the system prompt that you want to see the evaluation summary for.

  2. To create a red team evaluation using the original system prompt, click Run Red Team Evaluation. The Create Red Team Evaluation slide-out displays.

    System Prompt Evaluation Original Prompt Menu
  3. Enter a name for the evaluation.

  4. Select a target model from the drop-down menu.

    • Select a model that is similar to what you have in your environment to simulate the attacks against your system prompt.
    • Disclaimer: Models marked with a beta designation may be subject to lower usage quotas, limited availability, or ongoing development changes. As a result, these models may exhibit unexpected results, reduced performance, or intermittent failures during testing. Users should account for these limitations when selecting beta models.
  5. Optionally, click Advanced Options.

    • Select a project to apply runtime rulesets to interaction tagging.

      • If no project is selected, the default project is used.
    • Select an execution strategy.

      • Single: Runs each technique once per objective.
      • Random: Runs all techniques plus N additional random techniques.
      • Static prompt set: Uses a predefined set of static prompts for evaluation.
    • Set the maximum number of conversation turns allowed per technique when attempting to achieve an objective. The minimum is one and the maximum is five.

    • Set the number of independent sessions to run for each technique. The minimum is one and the maximum is five.

  6. Click Start Evaluation.

  7. Go to Security Testing > Red Teaming. You should see your red team evaluation in the table. A percentage complete and an estimated time to completion display for the evaluation.

  8. When the evaluation is complete, click the green arrow to view the details.

Analysis Questions

Expand the Analysis Questions to see the questions asked and the answers provided (by the user) when creating this evaluation.

Compiled System Prompt

The compiled system prompt is the enhanced version of the system prompt, with security improvements.

The compiled system prompt is the recommended system prompt to use.

Policies and Procedures

When using a compiled prompt, make sure you are following your organization's policies and procedures for system prompts.

Actions

Click the action button (three vertical dots) to view the Actions menu.

View Diff

Click View Diff to see a prompt comparison between the original and compiled prompts.

  • Removed content is highlighted in red.
  • Added content is highlighted in green.
  • Unchanged content is not highlighted.

Create Red Team Evaluation using Original System Prompt

  1. To view to a system prompt evaluation summary, select Security Testing > System Prompt, then click the green arrow for the system prompt that you want to see the evaluation summary for.

  2. To create a red team evaluation using the original system prompt, click Run Red Team Evaluation. The Create Red Team Evaluation slide-out displays.

    System Prompt Evaluation Compiled Prompt Menu
  3. Enter a name for the evaluation.

  4. Select a target model from the drop-down menu.

    • Select a model that is similar to what you have in your environment to simulate the attacks against your system prompt.
    • Note: A model with the beta tag indicates that it might have a lower quota or is a new model (from the provider) that could cause unexpected results or failures.
  5. Optionally, click Advanced Options.

    • Select a project to apply runtime rulesets to interaction tagging.

      • If no project is selected, the default project is used.
    • Select an execution strategy.

      • Single: Runs each technique once per objective.
      • Random: Runs all techniques plus N additional random techniques.
      • Static prompt set: Uses a predefined set of static prompts for evaluation.
    • Set the maximum number of conversation turns allowed per technique when attempting to achieve an objective. The minimum is one and the maximum is five.

    • Set the number of independent sessions to run for each technique. The minimum is one and the maximum is five.

  6. Click Start Evaluation.

  7. Go to Security Testing > Red Teaming. You should see your red team evaluation in the table. A percentage complete and an estimated time to completion display for the evaluation.

  8. When the evaluation is complete, click the green arrow to view the details.

Security Updates

Expand the Security Updates to see what improvements are achieved with the compiled prompt.

Usage Descriptions

UsageDescription
Total RequestsThe total number of requests for this evaluation.
Total TokensThe total number of tokens used for this evaluation, both prompt and completion tokens.
Prompt TokensThe number of prompt tokens used for this evaluation.
Completion TokensThe number of completion tokens used for this evaluation.