Environment Variables
Required Variables
CBL_API_KEY
Your Circuit Breaker Labs API key for authenticating with the evaluation service.Get your API key by contacting [email protected]
Optional Variables
CBL_API_BASE_URL
Custom base URL for the Circuit Breaker Labs API. Defaults towss://api.circuitbreakerlabs.ai/v1.
OPENAI_API_KEY
Required when using the OpenAI provider.OPENAI_BASE_URL
Custom base URL for OpenAI-compatible endpoints. Defaults tohttps://api.openai.com/v1.
OPENAI_ORG_ID
Optional OpenAI organization ID.OLLAMA_BASE_URL
Base URL for Ollama server. Defaults tohttp://localhost:11434.
Global Options
These flags can be used with any evaluation command and must be specified before the evaluation type.—log-level
Set the logging verbosity level.error, warn, info (default), debug, trace
—log-mode
Enable log mode to disable the TUI and output logs to stdout instead. Useful for CI/CD pipelines and debugging.—output-file
Specify a custom output file path for evaluation results. By default, results are saved to auto-generated files with timestamps:- Single-turn:
circuit_breaker_labs_single_turn_evaluation_YYYYMMDD_HHMMSS.json - Multi-turn:
circuit_breaker_labs_multi_turn_evaluation_YYYYMMDD_HHMMSS.json
—add-header
Add custom HTTP headers to provider requests. Can be specified multiple times for multiple headers."Key:Value"
Provider Configuration
OpenAI Provider
Required Options
Optional Parameters
Temperature and Sampling
Temperature and Sampling
—temperature
Sampling temperature between 0 and 2. Higher values make output more creative.—top-p
Nucleus sampling parameter. Alternative to temperature.—frequency-penalty
Number between -2.0 and 2.0. Positive values penalize repeated tokens based on frequency.—presence-penalty
Number between -2.0 and 2.0. Positive values penalize repeated tokens based on presence.
<float>Sampling temperature between 0 and 2. Higher values make output more creative.
<float>Nucleus sampling parameter. Alternative to temperature.
<float>Number between -2.0 and 2.0. Positive values penalize repeated tokens based on frequency.
<float>Number between -2.0 and 2.0. Positive values penalize repeated tokens based on presence.
Token Control
Token Control
—max-completion-tokens
Maximum number of tokens to generate in the completion.—stop
Up to 4 sequences where the API will stop generating tokens.—logit-bias
Modify likelihood of specified tokens appearing. Bias values between -100 and 100.
<integer>Maximum number of tokens to generate in the completion.
<string,string,...>Up to 4 sequences where the API will stop generating tokens.
<token_id:bias,...>Modify likelihood of specified tokens appearing. Bias values between -100 and 100.
Advanced Options
Advanced Options
—n
Number of chat completion choices to generate for each input.—logprobs
Return log probabilities of output tokens.—top-logprobs
Number between 0 and 20 specifying most likely tokens to return.—service-tier
Specifies the processing tier for the request.—reasoning-effort
Constrains effort on reasoning for reasoning models.—store
Whether to store the output of this chat completion request.
<integer>Number of chat completion choices to generate for each input.
<bool>Return log probabilities of output tokens.
<integer>Number between 0 and 20 specifying most likely tokens to return.
<auto|default|flex|scale|priority>Specifies the processing tier for the request.
<none|minimal|low|medium|high|xhigh>Constrains effort on reasoning for reasoning models.
<bool>Whether to store the output of this chat completion request.
Ollama Provider
Required Options
Optional Parameters
Model Options
Model Options
—temperature
Model temperature. Higher values = more creative (default: 0.8).—top-k
Reduces probability of generating nonsense. Higher = more diverse (default: 40).—top-p
Works with top-k. Higher values = more diverse text (default: 0.9).—num-predict
Maximum tokens to predict (default: 128, -1 = infinite, -2 = fill context).—num-ctx
Size of the context window (default: 2048).
<float>Model temperature. Higher values = more creative (default: 0.8).
<integer>Reduces probability of generating nonsense. Higher = more diverse (default: 40).
<float>Works with top-k. Higher values = more diverse text (default: 0.9).
<integer>Maximum tokens to predict (default: 128, -1 = infinite, -2 = fill context).
<integer>Size of the context window (default: 2048).
Sampling Control
Sampling Control
—mirostat
Enable Mirostat sampling (0 = disabled, 1 = Mirostat, 2 = Mirostat 2.0).—mirostat-eta
Mirostat learning rate (default: 0.1).—mirostat-tau
Controls balance between coherence and diversity (default: 5.0).—repeat-penalty
How strongly to penalize repetitions (default: 1.1).—repeat-last-n
How far back to look to prevent repetition (default: 64, 0 = disabled, -1 = num_ctx).—tfs-z
Tail free sampling - reduces impact of less probable tokens (default: 1).
<0|1|2>Enable Mirostat sampling (0 = disabled, 1 = Mirostat, 2 = Mirostat 2.0).
<float>Mirostat learning rate (default: 0.1).
<float>Controls balance between coherence and diversity (default: 5.0).
<float>How strongly to penalize repetitions (default: 1.1).
<integer>How far back to look to prevent repetition (default: 64, 0 = disabled, -1 = num_ctx).
<float>Tail free sampling - reduces impact of less probable tokens (default: 1).
Hardware and Performance
Hardware and Performance
—num-gpu
Number of layers to send to GPU(s).—num-thread
Number of threads to use during computation.—num-gqa
Number of GQA groups in transformer layer.
<integer>Number of layers to send to GPU(s).
<integer>Number of threads to use during computation.
<integer>Number of GQA groups in transformer layer.
Other Options
Other Options
—seed
Random number seed for generation (default: 0).—stop
Stop sequences (can be specified multiple times).—logprobs
Return log probabilities for each token.
<integer>Random number seed for generation (default: 0).
<string>Stop sequences (can be specified multiple times).
<bool>Return log probabilities for each token.
Custom Provider
For APIs that aren’t OpenAI-compatible, use Rhai scripts to translate between schemas.<string> (required)Endpoint URL to POST requests to. —script
<path> (required)Path to the Rhai script file that handles request/response translation.
See the examples/providers/ directory for example Rhai scripts.