Supported Providers
Page Assist includes built-in support for many popular providers:Cloud Providers
- OpenAI - GPT models (GPT-4, GPT-3.5, etc.)
- Anthropic (Claude) - Claude models
- Google AI - Gemini models
- Groq - Fast inference API
- DeepSeek - DeepSeek models
- Fireworks - Fast inference platform
- Together AI - Open source models
- OpenRouter - Unified API gateway
- Mistral - Mistral AI models
- xAI - Grok models
- Novita AI - AI inference platform
- Hugging Face - Inference API
- Infinigence AI - AI platform
- SiliconFlow - AI inference
- Moonshot - AI platform
- BigModel (Zhipu) - Chinese AI models
- MiniMax - Chinese AI models
- VolcEngine - ByteDance AI
- TencentCloud - Tencent AI
- AlibabaCloud - Alibaba AI
Self-Hosted / Local
- LLaMA.cpp - C++ implementation
- Llamafile - Single-file executables
- vLLM - High-performance inference
- Ollama (secondary instances)
- Vercel AI Gateway - AI proxy service
- Chutes - Custom deployments
Adding a Provider
Select Provider
Choose your provider from the dropdown:
- If listed, select it directly (e.g., “OpenAI”, “Anthropic”, “Groq”)
- If not listed, select “Custom”
Add Models (if required)
Some providers require manual model entry:
- Auto-detected: Ollama, LM Studio, LLaMA.cpp, Llamafile, vLLM
- Manual entry: Most cloud providers
gpt-4-turboclaude-3-5-sonnet-20241022llama-3.1-70b-versatile
Provider-Specific Setup
OpenAI
Get API Key
- Visit platform.openai.com
- Sign in or create an account
- Navigate to API Keys
- Create a new API key
Anthropic (Claude)
Get API Key
- Visit console.anthropic.com
- Sign in or create an account
- Navigate to API Keys
- Create a new API key
Groq
Get API Key
- Visit console.groq.com
- Sign in with your account
- Navigate to API Keys
- Create a new API key
LLaMA.cpp Server
vLLM
Custom Provider
For any unlisted OpenAI-compatible service:Enter Details
- Provider Name: Give it a recognizable name
- Base URL: Your API endpoint (must end with
/v1) - API Key: If required by your service
Model Management
Auto-Detection
The following providers support automatic model detection:- Ollama (secondary instances)
- LM Studio
- LLaMA.cpp
- Llamafile
- vLLM
Manual Model Entry
For most cloud providers, you need to manually add models:- After saving the provider, click “Add Model”
- Enter the exact model identifier
- Optionally specify if it’s a chat or embedding model
- Save the model
Model Types
Specify the model type for proper functionality:- Chat Models: For conversations (GPT-4, Claude, etc.)
- Embedding Models: For RAG features (text-embedding-ada-002, etc.)
Authentication
API Keys
Most cloud providers require API keys:Custom Headers
Some providers require custom headers:- In provider settings, expand “Advanced Options”
- Add custom headers as key-value pairs:
Bearer Tokens
For services using bearer tokens:- Enter the token in the “API Key” field
- Or add a custom header:
Authorization: Bearer your-token
Troubleshooting
Connection Failed
Models Not Loading
- Check if models need manual entry for your provider
- Verify the API endpoint supports
/v1/modelsfor auto-detection - Try manually adding model names
- Check API key permissions
Authentication Errors
- Verify API key format is correct
- Check for extra spaces in the API key
- Ensure the API key has necessary permissions
- Verify the API key hasn’t expired
Rate Limiting
If you encounter rate limit errors:- Wait before sending more requests
- Check your provider’s rate limits
- Consider upgrading your plan
- Use a different provider for high-volume usage
Best Practices
- Secure Your API Keys: Never share your API keys publicly
- Monitor Usage: Track API usage to avoid unexpected charges
- Test with Small Models: Start with smaller/cheaper models for testing
- Set Usage Limits: Configure spending limits in your provider’s dashboard
- Use Multiple Providers: Distribute usage across providers for reliability
- Keep Keys Rotated: Regularly rotate API keys for security
Provider Comparison
| Provider | Strengths | Best For |
|---|---|---|
| OpenAI | Most capable models | Complex tasks, reasoning |
| Anthropic | Long context, safety | Document analysis, safe AI |
| Groq | Ultra-fast inference | Speed-critical applications |
| Local (Ollama/LM Studio) | Privacy, no costs | Private data, development |
| Together AI | Open models, good pricing | Open source model access |
| OpenRouter | Access to many models | Model variety, fallbacks |
Next Steps
- Learn about Ollama setup
- Explore LM Studio integration
- Set up Chrome AI
- Configure Knowledge Base features