Overview
OpenCode supports multiple AI providers and models, allowing you to choose the best model for your use case and budget. You can configure different models for different agent types (coder, task, title).Supported providers
OpenCode supports the following AI providers (in order of auto-configuration preference):GitHub Copilot
Access multiple models through your GitHub Copilot subscription at no additional cost
Anthropic
Claude models with extended thinking and large context windows
OpenAI
GPT and o-series models with reasoning capabilities
Google Gemini
Gemini models with massive context windows
Groq
Fast inference for Llama and other open models
OpenRouter
Unified access to multiple providers
AWS Bedrock
Claude and other models on AWS infrastructure
Azure OpenAI
OpenAI models through Azure with enterprise features
Google Cloud Vertex AI
Gemini models on Google Cloud Platform
GitHub Copilot models
GitHub Copilot models are free with a GitHub Copilot subscription and automatically detected from your GitHub CLI credentials.
Authentication
OpenCode automatically detects GitHub Copilot credentials from:- GitHub CLI configuration (
~/.config/github-copilot/hosts.json) - Environment variable:
GITHUB_TOKEN
Available models
| Model ID | Name | Context Window | Max Tokens | Reasoning |
|---|---|---|---|---|
copilot.gpt-4o | GPT-4o | 128K | 16K | ✗ |
copilot.gpt-4o-mini | GPT-4o Mini | 128K | 4K | ✗ |
copilot.gpt-4.1 | GPT-4.1 | 128K | 16K | ✓ |
copilot.claude-3.5-sonnet | Claude 3.5 Sonnet | 90K | 8K | ✗ |
copilot.claude-3.7-sonnet | Claude 3.7 Sonnet | 200K | 16K | ✗ |
copilot.claude-sonnet-4 | Claude Sonnet 4 | 128K | 16K | ✗ |
copilot.o1 | OpenAI o1 | 200K | 100K | ✓ |
copilot.o3-mini | OpenAI o3-mini | 200K | 100K | ✓ |
copilot.o4-mini | OpenAI o4-mini | 128K | 16K | ✓ |
copilot.gemini-2.0-flash | Gemini 2.0 Flash | 1M | 8K | ✗ |
copilot.gemini-2.5-pro | Gemini 2.5 Pro | 128K | 64K | ✗ |
Anthropic (Claude)
Authentication
.opencode.json:
Available models
| Model ID | Name | Context Window | Max Tokens | Cost (per 1M tokens) | Reasoning |
|---|---|---|---|---|---|
claude-4-sonnet | Claude 4 Sonnet | 200K | 50K | 15.00 | ✓ |
claude-4-opus | Claude 4 Opus | 200K | 4K | 75.00 | ✗ |
claude-3.7-sonnet | Claude 3.7 Sonnet | 200K | 50K | 15.00 | ✓ |
claude-3.5-sonnet | Claude 3.5 Sonnet | 200K | 5K | 15.00 | ✗ |
claude-3.5-haiku | Claude 3.5 Haiku | 200K | 4K | 4.00 | ✗ |
claude-3-opus | Claude 3 Opus | 200K | 4K | 75.00 | ✗ |
claude-3-haiku | Claude 3 Haiku | 200K | 4K | 1.25 | ✗ |
OpenAI
Authentication
.opencode.json:
Available models
| Model ID | Name | Context Window | Max Tokens | Cost (per 1M tokens) | Reasoning |
|---|---|---|---|---|---|
gpt-4.1 | GPT-4.1 | 1M | 20K | 8.00 | ✗ |
gpt-4.1-mini | GPT-4.1 Mini | 200K | 20K | 1.60 | ✗ |
gpt-4.1-nano | GPT-4.1 Nano | 1M | 20K | 0.40 | ✗ |
gpt-4.5-preview | GPT-4.5 Preview | 128K | 15K | 150.00 | ✗ |
gpt-4o | GPT-4o | 128K | 4K | 10.00 | ✗ |
gpt-4o-mini | GPT-4o Mini | 128K | — | 0.60 | ✗ |
o1 | O1 | 200K | 50K | 60.00 | ✓ |
o1-pro | O1 Pro | 200K | 50K | 600.00 | ✓ |
o1-mini | O1 Mini | 128K | 50K | 4.40 | ✓ |
o3 | O3 | 200K | — | 40.00 | ✓ |
o3-mini | O3 Mini | 200K | 50K | 4.40 | ✓ |
o4-mini | O4 Mini | 128K | 50K | 4.40 | ✓ |
Google Gemini
Authentication
.opencode.json:
Available models
| Model ID | Name | Context Window | Max Tokens | Cost (per 1M tokens) |
|---|---|---|---|---|
gemini-2.5 | Gemini 2.5 Pro | 1M | 50K | 10.00 |
gemini-2.5-flash | Gemini 2.5 Flash | 1M | 50K | 0.60 |
gemini-2.0-flash | Gemini 2.0 Flash | 1M | 6K | 0.40 |
gemini-2.0-flash-lite | Gemini 2.0 Flash Lite | 1M | 6K | 0.30 |
Groq
Authentication
.opencode.json:
Available models
| Model ID | Name | Context Window | Max Tokens | Cost (per 1M tokens) | Reasoning |
|---|---|---|---|---|---|
qwen-qwq | Qwen QwQ | 128K | 50K | 0.39 | ✗ |
llama-3.3-70b-versatile | Llama 3.3 70B | 128K | — | 0.79 | ✗ |
meta-llama/llama-4-scout-17b-16e-instruct | Llama 4 Scout | 128K | — | 0.34 | ✗ |
meta-llama/llama-4-maverick-17b-128e-instruct | Llama 4 Maverick | 128K | — | 0.20 | ✗ |
deepseek-r1-distill-llama-70b | DeepSeek R1 Distill | 128K | — | 0.99 | ✓ |
OpenRouter
Authentication
.opencode.json:
Available models
OpenRouter provides access to models from multiple providers with unified billing: OpenAI models:openrouter.gpt-4.1,openrouter.gpt-4.1-mini,openrouter.gpt-4.1-nanoopenrouter.gpt-4.5-preview,openrouter.gpt-4o,openrouter.gpt-4o-miniopenrouter.o1,openrouter.o1-pro,openrouter.o1-miniopenrouter.o3,openrouter.o3-mini,openrouter.o4-mini
openrouter.claude-3.5-sonnet,openrouter.claude-3.7-sonnetopenrouter.claude-3.5-haiku,openrouter.claude-3-haikuopenrouter.claude-3-opus
openrouter.gemini-2.5,openrouter.gemini-2.5-flash
openrouter.deepseek-r1-free- Free DeepSeek R1 (163K context)
See OpenRouter pricing for current rates.
AWS Bedrock
Authentication
Bedrock uses standard AWS credentials:Available models
| Model ID | Name | Context Window | Cost (per 1M tokens) | Reasoning |
|---|---|---|---|---|
bedrock.claude-3.7-sonnet | Claude 3.7 Sonnet | — | 15.00 | ✗ |
Azure OpenAI
Authentication
.opencode.json:
Available models
Azure OpenAI provides the same models as OpenAI with theazure. prefix:
azure.gpt-4.1,azure.gpt-4.1-mini,azure.gpt-4.1-nanoazure.gpt-4.5-preview,azure.gpt-4o,azure.gpt-4o-miniazure.o1,azure.o1-miniazure.o3,azure.o3-mini,azure.o4-mini
Pricing matches OpenAI’s rates. See Azure OpenAI pricing.
Google Cloud Vertex AI
Authentication
Available models
| Model ID | Name | Context Window | Max Tokens |
|---|---|---|---|
vertexai.gemini-2.5 | Gemini 2.5 Pro | 1M | 50K |
vertexai.gemini-2.5-flash | Gemini 2.5 Flash | 1M | 50K |
xAI (Grok)
Authentication
Available models
| Model ID | Name | Context Window | Max Tokens | Cost (per 1M tokens) |
|---|---|---|---|---|
grok-3-beta | Grok 3 Beta | 131K | 20K | 15.00 |
grok-3-mini-beta | Grok 3 Mini Beta | 131K | 20K | 0.50 |
grok-3-fast-beta | Grok 3 Fast Beta | 131K | 20K | 25.00 |
grok-3-mini-fast-beta | Grok 3 Mini Fast Beta | 131K | 20K | 4.00 |
Choosing the right model
For coding (coder agent)
- Best quality
- Balanced
- Fast & affordable
claude-4-sonnet- Best overall for complex reasoningo1oro3- Excellent for algorithmic challengesgpt-4.1- Strong general coding capabilities
For code search (task agent)
Fast, cost-effective models work best:gpt-4.1-minigpt-4o-miniclaude-3.5-haikugemini-2.5-flash
For titles (title agent)
Any fast model works (auto-configured, maxTokens=80):gpt-4o-miniclaude-3.5-haikugemini-2.0-flash-lite
Model capabilities
Reasoning models
Models with extended thinking/reasoning capabilities:- OpenAI: o1, o1-pro, o1-mini, o3, o3-mini, o4-mini
- Anthropic: claude-4-sonnet, claude-3.7-sonnet
- Groq: deepseek-r1-distill-llama-70b
reasoningEffort (low, medium, high):
File attachments
Most models support file attachments except:o3-minillama-3.3-70b-versatile- Groq’s Qwen and DeepSeek models
Context windows
Largest context windows:- 1M tokens: Gemini models, VertexAI Gemini
- 1M+ tokens:
gpt-4.1,gpt-4.1-nano(1,047,576 tokens) - 200K tokens: Claude models, OpenAI o-series,
gpt-4.1-mini
Cost optimization
Prompt caching
Many models support prompt caching to reduce costs on repeated context:- Anthropic: All Claude models
- OpenAI: GPT-4.1, o-series models
- Gemini: Not yet supported
Strategy recommendations
Budget-conscious setup
Budget-conscious setup
GitHub Copilot (free)
GitHub Copilot (free)
Performance-optimized setup
Performance-optimized setup
Related topics
- Configuration - Configure models and providers
- Architecture - How OpenCode uses AI models internally