Default models
Docbot ships with these default model configurations:All models are accessed through Vercel AI Gateway, which requires an
AI_GATEWAY_API_KEY environment variable.Model types and their roles
Planning models
planning - Used for general task planning and strategy
- Default:
openai/gpt-5.2 - Purpose: Analyzing documentation needs and creating structured plans
- Characteristics: Balanced speed and reasoning ability
planningHeavy - Used for complex, multi-step planning
- Default:
anthropic/claude-opus-4.5 - Purpose: Deep analysis and sophisticated planning tasks
- Characteristics: Highest quality reasoning, slower and more expensive
Writing models
prose - Used for writing and editing documentation
- Default:
anthropic/claude-sonnet-4.5 - Purpose: Generating high-quality documentation prose
- Characteristics: Excellent writing quality, context awareness
Utility models
fast - Used for quick, simple tasks
- Default:
openai/gpt-5.2 - Purpose: Fast responses for straightforward operations
- Characteristics: Low latency, cost-effective
nano - Used for very simple, high-volume tasks
- Default:
google/gemini-3-flash - Purpose: Minimal processing tasks
- Characteristics: Fastest, cheapest option
context - Used for large context window needs
- Default:
google/gemini-3-pro-preview - Purpose: Processing large documents or codebases
- Characteristics: Extended context length
Embedding models
embedding - Default embedding model for indexing
- Default:
openai/text-embedding-3-small - Vector size: 1536 dimensions
- Purpose: General-purpose semantic search
embeddingLarge - High-quality embeddings
- Default:
openai/text-embedding-3-large - Purpose: When higher accuracy is needed
- Characteristics: Better quality, more expensive
Customizing models
Use the correct format
Model identifiers must follow the format:Examples:
openai/gpt-5.2anthropic/claude-sonnet-4.5google/gemini-3-flashcohere/command-r-plus
Model selection strategies
Cost optimization
To minimize costs, use smaller/faster models:Quality optimization
For best output quality, use premium models:Balanced approach
Use premium models only where they matter most:Provider-specific notes
OpenAI models
Supported models:- GPT-5.2 (latest)
- GPT-4.5 series
- Embedding models (text-embedding-3-small, text-embedding-3-large)
Anthropic models
Claude model tiers:- Opus: Highest intelligence, most expensive
- Sonnet: Balanced performance and cost
- Haiku: Fastest, most cost-effective
Google models
Gemini variants:- Pro: Large context windows
- Flash: Optimized for speed
Cohere models
Docbot also supports Cohere for:- Reranking:
cohere/rerank-v3.5 - Generation models as needed
Embedding model considerations
Vector dimensions
The default Qdrant configuration expects 1536-dimensional vectors. If you change embedding models, ensure:- The new model uses 1536 dimensions, or
- You rebuild your Qdrant collections with the correct vector size
Changing embedding models
If you switch embedding models after indexing:Future provider support
Docbot currently requires Vercel AI Gateway viaAI_GATEWAY_API_KEY. Support for additional providers is planned:
- Direct API access (OpenAI, Anthropic, etc.)
- Azure OpenAI
- Custom/self-hosted models
- Local models via Ollama
Validating model configuration
The configuration schema enforces:Next steps
- Learn about local development setup
- Understand Qdrant configuration
- Review troubleshooting tips for model-related issues