Configuring AI providers
Claude Octopus orchestrates up to three AI providers: Codex (OpenAI), Gemini (Google), and Claude (Anthropic). This guide covers authentication, cost awareness, and provider selection.Provider roles
Each provider has a distinct role in multi-AI workflows:Codex (OpenAI)
🔴 Implementation depth
- Code patterns and technical analysis
- Architecture decisions
- Framework-specific guidance
- Debugging assistance
Gemini (Google)
🟡 Ecosystem breadth
- Alternatives and market trends
- Security review perspectives
- Research synthesis
- Community insights
Claude (Anthropic)
🔵 Orchestration
- Quality gates and validation
- Consensus building
- Final synthesis
- Strategic guidance
Provider authentication
Claude Octopus supports two authentication methods: OAuth (recommended) and API keys.OAuth (recommended)
OAuth uses your existing AI subscriptions—no per-token billing.- Codex
- Gemini
- Claude
Authentication:What you need:Installation:
- Active ChatGPT subscription (Plus, Team, or Enterprise)
- Codex CLI installed
- Included in ChatGPT subscription
- No additional per-token charges
API keys
API keys use per-token billing—you pay for each API call.- Codex
- Gemini
- Claude
Set environment variable:Get API key:
- Visit platform.openai.com/api-keys
- Create new secret key
- Add to environment variables
- GPT-5.3-Codex: 14 per MTok input/output
- Estimated: ~$0.02-0.10 per query
What works with just Claude
Claude Octopus works with zero external providers. You get every feature on day one: Single-provider features:- ✅ All 33 personas
- ✅ Structured workflows (discover, define, develop, deliver)
- ✅ Smart routing (
/octocommand) - ✅ Context detection
- ✅ Every skill
- ✅ Quality gates
- ✅ Command composition
- Parallel research (Codex + Gemini + Claude)
- Adversarial debate (three-way)
- Cross-model review (consensus building)
- Provider-specific strengths (implementation depth + ecosystem breadth)
Running without external providers
- Research phase uses Claude only (no parallel multi-AI research)
- Consensus building uses Claude reasoning (no cross-model debate)
- Review uses Claude quality gates (no adversarial cross-checking)
Adding Codex and Gemini
Add external providers to unlock multi-AI orchestration.Cost awareness and provider selection
Claude Octopus provides full cost transparency with visual indicators and smart routing.Visual indicators
When multi-AI workflows run, you’ll see exactly which providers are active:| Indicator | Provider | Cost Source |
|---|---|---|
| 🔴 | Codex CLI | Your OPENAI_API_KEY or ChatGPT subscription |
| 🟡 | Gemini CLI | Your GEMINI_API_KEY or Google AI subscription |
| 🔵 | Claude | Included with Claude Code subscription |
| 🐙 | Multi-AI mode | Multiple providers active |
Cost comparison
Codex (OpenAI)
Codex (OpenAI)
Models:
- GPT-5.3-Codex: 14 per MTok input/output
- Spark: 15x faster than standard
- Mini: ~1.25 per MTok
- Simple: ~$0.01-0.02
- Complex: ~$0.05-0.10
- With context: ~$0.10-0.15
- Technical implementation analysis
- Code generation
- Framework-specific guidance
Gemini (Google)
Gemini (Google)
Models:
- Gemini Pro: Standard model
- Flash: Faster, cheaper for simple queries
- Simple: ~$0.01
- Complex: ~$0.02-0.03
- Ecosystem research
- Market trends
- Alternative approaches
Claude (Anthropic)
Claude (Anthropic)
Models:
- Sonnet 4.6: Included with Claude Code subscription
- Opus 4.6: 25 per MTok input/output
- Opus 4.6 Fast: 150 per MTok (6x cost, lower latency)
- Sonnet: Included (no additional cost)
- Opus: Extra-usage billing
- Fast Opus: 6x more expensive (use sparingly)
- Orchestration and synthesis
- Quality gates
- Consensus building
Smart cost routing
Claude Octopus routes to the cheapest provider that can handle the job. Routing logic:- Simple tasks → Claude Sonnet (free with subscription)
- Technical implementation → Codex if available, else Claude Opus
- Ecosystem research → Gemini if available, else Claude Sonnet
- Complex reasoning → Claude Opus or Codex depending on task
- Interactive queries → Fast mode only when user is actively waiting
Fast Opus mode (Claude Code v2.1.36+)
When fast mode is used:- Interactive single-shot Opus queries where user is actively waiting
- Never in autonomous/background mode (no human waiting = no latency benefit)
- Standard mode for all multi-phase workflows
- Fast mode only for interactive queries
Cost estimation
Typical workflow costs:| Workflow | Providers | Estimated Cost |
|---|---|---|
/octo:discover | Claude only | $0 (included) |
/octo:discover | Codex + Gemini + Claude | ~$0.05-0.15 |
/octo:debate | Codex + Gemini + Claude | ~$0.10-0.30 |
/octo:embrace | Claude only | $0 (included) |
/octo:embrace | Codex + Gemini + Claude | ~$0.30-0.80 |
/octo:factory | Codex + Gemini + Claude | ~$0.50-2.00 |
Provider configuration reference
Environment variables
Provider detection
Check which providers are available:Next steps
Using commands
Learn command structure, the smart router, and the top 8 commands
Working with workflows
Understand workflow progression, quality gates, and autonomy modes
Configuration
Environment variables, settings, and custom hooks
Visual indicators
Deep dive into cost transparency and provider status
