Common Error Scenarios
Action fails with 'Error: 401' or authentication error
Action fails with 'Error: 401' or authentication error
Symptom:Cause: Invalid or missing Circuit Breaker Labs API key.Solutions:
-
Verify the API key is set as a secret:
- Go to your repository Settings → Secrets and variables → Actions
- Confirm
CBL_API_KEY(or your chosen secret name) exists - If missing, add it with your valid API key from Circuit Breaker Labs
-
Check the secret reference in your workflow:
Ensure the secret name matches exactly (case-sensitive).
-
Verify your API key is valid:
- Test your API key with a direct cURL request:
- If this fails, your API key may be expired or invalid. Contact [email protected].
API keys are environment-specific. Ensure you’re using the correct key for your account.
Action fails with 'Error: 422' or validation error
Action fails with 'Error: 422' or validation error
Symptom:Cause: Invalid input parameters sent to the API.Common Issues:
-
Threshold values out of range:
- Thresholds must be between 0.0 and 1.0
-
Invalid model name:
- Verify the model name exists on OpenRouter or OpenAI
-
Multi-turn max-turns must be even:
-
Missing required parameters:
- Review the action reference for all required inputs
- Ensure all required fields are provided
-
Enable workflow debug logging:
- Go to Settings → Secrets and variables → Actions
- Add
ACTIONS_STEP_DEBUGwith valuetrue - Re-run the workflow to see detailed parameter values
-
Check the API response for specific validation errors:
Workflow fails due to threshold exceeded
Workflow fails due to threshold exceeded
Symptom:Cause: The evaluation detected unsafe behavior exceeding your configured
fail-action-threshold.This is expected behavior—not an error! The action is designed to fail your workflow when safety issues are detected.Next Steps:-
Review the failed test cases:
- Examine the
User InputandModel Responsefor each failure - Identify patterns in the types of attacks that succeeded
- Note the
Safety Scorefor each failed case
- Examine the
-
Improve your system prompt:
- Add explicit safety instructions
- Include refusal examples
- Strengthen boundary conditions
- See configuration guide for prompt engineering tips
-
Adjust thresholds if needed:
fail-case-threshold: Lower values are more strict (fewer cases marked as failures)fail-action-threshold: Controls overall failure rate tolerance
-
Iterate and re-test:
- Update your system prompt
- Commit changes to trigger a new evaluation
- Compare results to previous runs
Action times out or hangs
Action times out or hangs
Symptom:
- Action runs for an extended period without completing
- GitHub Actions job timeout (default 6 hours)
-
Large number of test variations:
- High
variationsvalues increase test count exponentially - High
maximum-iteration-layersincreases complexity
- High
-
Slow model responses:
- Some models take longer to respond
- API rate limits may slow execution
-
Reduce test complexity for faster feedback:
-
Set a reasonable job timeout:
-
Use faster models for development:
- Switch to faster models during development
- Use more thorough testing in production
-
Split test case groups:
OpenAI fine-tune actions fail with 'Model not found'
OpenAI fine-tune actions fail with 'Model not found'
Symptom:Cause: Invalid OpenAI fine-tuned model name or incorrect API key.Solutions:
-
Verify the model name format:
-
Check OpenAI API key permissions:
- The API key must have access to the specific fine-tuned model
- Verify the key belongs to the correct organization
- Test access with OpenAI CLI:
-
Ensure the model exists:
- List your fine-tuned models:
- Confirm the model status is “succeeded”
-
Store OpenAI API key as a secret:
Network errors or connection timeouts
Network errors or connection timeouts
Symptom:Causes:
- Network connectivity issues
- API service downtime
- GitHub Actions runner network restrictions
-
Check Circuit Breaker Labs API status:
- Visit https://api.circuitbreakerlabs.ai/v1/docs
- If the page loads, the API is likely operational
-
Add retry logic to your workflow:
-
Contact support if issues persist:
- Email: [email protected]
- Include workflow run ID and timestamp
Interpreting Failure Output
When evaluations fail due to threshold violations, the action outputs detailed information about each failed test case.Single-Turn Failure Output
- Overall Failure Rate: Percentage of total test cases that failed (compared against
fail-action-threshold) - Layer: Iteration layer number (0 is the base attack, higher numbers are refined attacks)
- Safety Score: Confidence that the response is safe (0.0 = definitely unsafe, 1.0 = definitely safe)
- Scores below
fail-case-thresholdare marked as failures
- Scores below
- User Input: The adversarial input that was sent to the model
- Model Response: How the model actually responded (what made it fail)
Multi-Turn Failure Output
- Overall Failure Rate: Percentage of conversations that failed
- Case: Sequential conversation number
- Min Safety Score: Lowest safety score across all turns in the conversation
- Conversation: Complete turn-by-turn dialogue
- Each turn shows role (user/assistant) and content
- Helps identify at which point the model was compromised
Lower safety scores indicate higher confidence that the response is unsafe. A score of 0.15 means there’s an 85% confidence the response violates safety guidelines.
Debugging Workflow Issues
Enable Detailed Logging
Add these secrets to see more detailed output:Check Action Version
Ensure you’re using a stable version tag:Validate Input Syntax
Use YAML validators to check syntax:Test Locally with Docker
You can test the underlying Python scripts locally:Getting Help
If you continue to experience issues:-
Check existing GitHub Issues:
- Circuit Breaker Labs Actions Issues
- Search for similar problems and solutions
-
Open a new issue:
- Include your workflow YAML (remove secrets)
- Provide the full error message and workflow run log
- Specify action version and GitHub runner OS
-
Contact support:
- Email: [email protected]
- Include workflow run ID and detailed description
- Review documentation: