Overview
vLLora acts as a unified gateway to multiple AI providers. Each provider requires its own API credentials. This guide covers how to configure and manage API keys for different providers.Supported Providers
vLLora supports the following AI providers out of the box:Tier 1 Providers (Built-in)
OpenAI
GPT-4, GPT-4o, GPT-4o-mini, and GPT-3.5 models
Anthropic
Claude 3.5 Sonnet, Claude 3 Opus, and Haiku models
Google Gemini
Gemini Pro and Gemini Flash models
Amazon Bedrock
Access to multiple foundation models via AWS
Additional Providers
- DeepSeek - Advanced AI models
- xAI (Grok) - X.ai models
- DeepInfra - Cloud-hosted open models
- Fireworks AI - Fast inference for compound AI
- Together AI - Open-source model platform
- Azure OpenAI - Microsoft Azure AI services
- Groq - LPU-powered ultra-fast inference
- Mistral AI - European AI leader
- OpenRouter - Multi-model routing
- Parasail - Cost-efficient AI deployment
- Vertex AI - Google Cloud AI platform
- Z.AI (Zhipu) - Chinese LLM provider
Configuring API Keys
Via Web UI (Recommended)
-
Start vLLora:
-
Open the UI at
http://localhost:9091 - Navigate to Providers or Settings
- Select a provider and click Configure
- Enter your API key and save
Via REST API
You can also manage API keys programmatically:List Providers
Add/Update Provider Credentials
Delete Provider Credentials
Provider-Specific Configuration
OpenAI
Get an API key: platform.openai.com/api-keysgpt-4o- Latest GPT-4 Omni modelgpt-4o-mini- Faster, cheaper GPT-4ogpt-4-turbo- Previous generation GPT-4gpt-3.5-turbo- Fast and cost-effective
Anthropic
Get an API key: console.anthropic.comclaude-3-5-sonnet-20241022- Most capable Claude modelclaude-3-opus-20240229- Highest intelligenceclaude-3-sonnet-20240229- Balanced performanceclaude-3-haiku-20240307- Fastest responses
Google Gemini
Get an API key: aistudio.google.com/apikeygemini-2.0-flash-exp- Latest experimental flashgemini-1.5-pro- Powerful and versatilegemini-1.5-flash- Fast and efficient
Amazon Bedrock
AWS credentials required:anthropic.claude-3-5-sonnet-20241022-v2:0anthropic.claude-3-opus-20240229-v1:0meta.llama3-70b-instruct-v1:0- And many more via AWS Bedrock
Azure OpenAI
Azure-specific configuration:Vertex AI (Google Cloud)
Service account required:OpenAI-Compatible Providers
Many providers (DeepSeek, Groq, Fireworks, etc.) use the OpenAI API format:Custom Providers
You can add custom OpenAI-compatible endpoints:Create Custom Provider
openai- OpenAI-compatible APIanthropic- Anthropic-compatible APIgemini- Google Gemini APIbedrock- AWS Bedrock API
Add Credentials to Custom Provider
Delete Custom Provider
Multi-Project Support
vLLora supports multiple projects, each with its own API keys:Security Best Practices
Storage
- API keys are stored in the local SQLite database at
~/.vllora/vllora.db - Keys are associated with projects and stored as encrypted credentials
- The database is only accessible to the user running vLLora
Access Control
- vLLora runs locally and does not expose credentials over the network
- API keys are never logged or included in traces
- Use project-based separation for different environments
Key Rotation
- Generate a new API key from your provider
- Update the key in vLLora (via UI or API)
- Test that requests work with the new key
- Revoke the old key from the provider’s dashboard
Environment Separation
For production deployments:Syncing Providers
vLLora can sync provider definitions from the LangDB API:- Provider metadata (descriptions, endpoints, etc.)
- New providers added to LangDB
- Deactivates providers removed from LangDB
Syncing providers does not affect your configured API keys. It only updates provider definitions.
Troubleshooting
Invalid API Key
Provider Not Found
vllora sync --providers to update provider definitions.
Missing Credentials
Rate Limiting
vLLora respects provider rate limits. If you hit rate limits:- Configure multiple providers with fallback
- Use vLLora’s built-in rate limiting middleware
- Upgrade your provider plan for higher limits
Next Steps
Models
Learn about model configuration
Chat Completions
Make your first API request
Monitoring
Monitor API usage and costs
Projects
Manage multiple projects