Common Parameters
These parameters are available across all Circuit Breaker Labs actions.Threshold Configuration
The failure rate threshold that determines whether the GitHub Action workflow fails. If the test failure rate exceeds this value, the action exits with a non-zero status code.Value range:
0.0 to 1.0 (representing 0% to 100%)Example: 0.80 means the action fails if more than 80% of test cases fail.See Thresholds for detailed examples.The safety score threshold for determining whether an individual test case passes or fails. Test cases with safety scores below this threshold are considered failures.Value range:
0.0 to 1.0Example: 0.5 means any test case with a safety score below 0.5 is marked as failed.See Thresholds for detailed examples.Authentication
Your Circuit Breaker Labs API key for authenticating requests to the evaluation API.
Test Filtering
Optional space-separated list of test case groups to run. When omitted, all available test groups are executed.This parameter allows you to target specific categories of security tests, enabling focused evaluation during development or CI/CD.See Test Case Groups for available groups and usage examples.
Single-Turn Evaluation Parameters
These parameters apply tosingleturn-evaluate-system-prompt and singleturn-evaluate-openai-finetune actions.
Evaluation Configuration
Number of test variations to generate and run for each test case. Higher values provide more comprehensive coverage but increase evaluation time and API usage.Minimum:
1Example: 3 generates three variations of each test case.Maximum number of iteration layers for adversarial testing. Each layer represents a level of increasingly sophisticated attack attempts.Minimum:
1Example: 2 runs tests across two iteration layers.Higher iteration layers test your model’s resilience against more advanced jailbreak attempts.
Model Configuration (System Prompt)
The system prompt text to evaluate. This is the instruction text that guides your model’s behavior.Available in:
singleturn-evaluate-system-prompt, multiturn-evaluate-system-promptThe OpenRouter model identifier to use for evaluation. Must be a valid model name supported by OpenRouter.Available in:
singleturn-evaluate-system-prompt, multiturn-evaluate-system-promptExamples:anthropic/claude-3.7-sonnetopenai/gpt-4meta-llama/llama-3.1-70b-instruct
Model Configuration (Fine-Tune)
The fully qualified name of your OpenAI fine-tuned model to evaluate.Available in:
singleturn-evaluate-openai-finetune, multiturn-evaluate-openai-finetuneFormat: ft:gpt-4-0125-preview:organization:model-name:identifierYour OpenAI API key for accessing fine-tuned models.Available in:
singleturn-evaluate-openai-finetune, multiturn-evaluate-openai-finetuneMulti-Turn Evaluation Parameters
These parameters apply tomultiturn-evaluate-system-prompt and multiturn-evaluate-openai-finetune actions.
Maximum number of conversation turns to simulate in multi-turn evaluations. Must be an even number.Constraint: Must be an even integerExample:
4 simulates conversations with up to 4 turns (2 user messages, 2 assistant responses)Multi-turn evaluations test whether adversarial prompts can exploit your model across extended conversations.
Space-separated list of multi-turn test types to execute. Each test type represents a different attack strategy.Format: Multiple values separated by spaces
Complete Examples
Single-Turn System Prompt Evaluation
Multi-Turn Fine-Tune Evaluation
Parameter Validation
The actions perform validation on input parameters:- Numeric ranges: Threshold values must be between 0.0 and 1.0
- Integer constraints:
max-turnsmust be an even number - Required combinations: Fine-tune actions require both
model-nameandopenai-api-key - Format validation:
test-typesandtest-case-groupsaccept space-separated values