Provider List
OpenFang supports these providers out of the box:Anthropic
Claude Opus/Sonnet/Haiku
OpenAI
GPT-4o, GPT-4, o1, o3
Gemini 2.0 Flash/Pro
Groq
Ultra-fast inference
DeepSeek
Coding specialists
xAI
Grok models
OpenRouter
Multi-provider routing
Mistral
European open models
Together AI
Open source models
Fireworks
Fast open models
Cohere
Command R+
Perplexity
Search-augmented LLMs
AI21 Labs
Jamba models
Cerebras
Fastest inference
SambaNova
Enterprise inference
Hugging Face
70k+ models
Replicate
Cloud model hosting
Ollama
Local models
vLLM
Self-hosted inference
LM Studio
Desktop inference
Qwen
Alibaba models
MiniMax
Chinese frontier
Zhipu AI
GLM models
Moonshot
Kimi long context
Qianfan
Baidu ERNIE
Bedrock
AWS hosted models
GitHub Copilot
Copilot API
Quick Setup
1. Anthropic (Claude)
claude-opus-4-20250514, claude-sonnet-4-20250514, claude-haiku-4.5-20251001, claude-opus-4.6, claude-sonnet-4.6
API Key: Get from console.anthropic.com
2. OpenAI
gpt-4o, gpt-4o-mini, gpt-4-turbo, gpt-4, o1, o1-mini, o3-mini, gpt-3.5-turbo
API Key: Get from platform.openai.com
3. Google Gemini
gemini-2.0-flash-exp, gemini-2.0-flash-thinking-exp, gemini-1.5-pro, gemini-1.5-flash, gemini-exp-1206
API Key: Get from aistudio.google.com
4. Groq
llama-3.3-70b-versatile, llama-3.1-70b-versatile, llama-3.2-90b-vision-preview, mixtral-8x7b-32768
API Key: Get from console.groq.com
Free Tier: 30 requests/minute, 7,000 requests/day
5. DeepSeek
deepseek-chat, deepseek-reasoner, deepseek-coder
API Key: Get from platform.deepseek.com
6. Ollama (Local)
llama3.2, mistral, codellama, qwen2.5, etc.
7. OpenRouter
8. xAI (Grok)
grok-beta, grok-vision-beta
API Key: Get from console.x.ai
Chinese Providers
Qwen (Alibaba)
qwen-max, qwen-plus, qwen-turbo, qwen3-235b-a22b, qwen-vl-plus, qwen-coder-plus
Zhipu AI (GLM)
glm-4-plus, glm-4-flash, glm-5-20250605, glm-4v-plus, codegeex-4
MiniMax
MiniMax-M2.5, MiniMax-M2.1, minimax-text-01, abab6.5-chat
Context: Up to 1M tokens
Moonshot (Kimi)
moonshot-v1-128k, moonshot-v1-32k, moonshot-v1-8k
Baidu Qianfan (ERNIE)
ernie-4.5-8k, ernie-4.0-turbo-8k, ernie-speed-128k
Self-Hosted & Local
vLLM
LM Studio
- Download LM Studio
- Load a model (e.g., Llama 3.3 70B)
- Start local server (default port 1234)
- Configure OpenFang to use it
Enterprise & Cloud
AWS Bedrock
bedrock/anthropic.claude-opus-4-6, bedrock/anthropic.claude-sonnet-4-6, bedrock/amazon.nova-pro-v1:0, bedrock/meta.llama3-3-70b-instruct-v1:0
GitHub Copilot
Additional Providers
Mistral AI
mistral-large, mistral-medium, mistral-small, codestral
Together AI
Fireworks AI
Cohere
command-r-plus, command-r, command
Perplexity
sonar, sonar-pro - search-augmented responses
AI21 Labs
jamba-1.5-large, jamba-1.5-mini
Cerebras
SambaNova
Hugging Face
Replicate
Custom Provider URLs
Override any provider’s base URL:Provider Status Check
Check which providers are configured:Multi-Provider Strategy
Troubleshooting
Authentication Errors
Rate Limits
Configure fallback providers to handle rate limits:Model Not Found
Next Steps
Model Configuration
Configure model routing and fallbacks
Channel Setup
Connect messaging platforms