Supported LLM Providers
ScrapeGraphAI supports 20+ LLM providers, giving you flexibility to choose the best model for your scraping needs. Each provider offers different models, pricing, and performance characteristics.Quick Comparison
OpenAI
Industry-leading models with GPT-4o and GPT-4o-mini. Best for complex scraping tasks.
Ollama
Run models locally with Llama 3.2, Mistral, and more. Free and private.
Azure OpenAI
Enterprise-grade OpenAI models with Azure infrastructure.
Google Gemini
Powerful Gemini 2.0 Pro with 2M token context window.
Groq
Ultra-fast inference with Llama and Gemma models.
Anthropic Claude
Claude 3.5 Sonnet and Opus for advanced reasoning.
All Supported Providers
Cloud Providers
| Provider | Popular Models | Context Window | Best For |
|---|---|---|---|
| OpenAI | gpt-4o, gpt-4o-mini | 128K tokens | Complex scraping, best accuracy |
| Anthropic | Claude 3.5 Sonnet, Claude Opus | 200K tokens | Advanced reasoning, long content |
| Google Gemini | Gemini 2.0 Pro, Flash | 1M-2M tokens | Massive context, multimodal |
| Azure OpenAI | gpt-4o, gpt-4-turbo | 128K tokens | Enterprise deployments |
| Groq | Llama 3.3 70B, Gemma 2 | 128K tokens | Speed, cost-effective |
| Mistral AI | Mistral Large, Codestral | 128K tokens | European hosting, coding |
| Deepseek | Deepseek-V3, R1 | 128K tokens | Cost-effective, reasoning |
| Together AI | Llama 3.1 405B, Mixtral | 128K tokens | Open models, flexible |
| Fireworks | Llama 3.1, Mixtral | 131K tokens | Fast inference |
| NVIDIA NIM | Llama 3.3, Nemotron | 128K tokens | GPU-optimized |
| AWS Bedrock | Claude, Llama, Mistral | Up to 200K | AWS ecosystem |
| xAI | Grok-3, Grok-3 Mini | 1M tokens | Latest models |
Local/Self-Hosted
| Provider | Description | Best For |
|---|---|---|
| Ollama | Run models locally (Llama, Mistral, Gemma) | Privacy, no API costs |
| Hugging Face | 300+ open models | Research, custom models |
| OneAPI | Unified API for Chinese models | Chinese content |
Enterprise Solutions
| Provider | Description | Best For |
|---|---|---|
| Azure OpenAI | Microsoft-hosted OpenAI | Enterprise compliance |
| AWS Bedrock | Serverless foundation models | AWS infrastructure |
| Google Vertex AI | Google Cloud AI platform | GCP ecosystem |
| Clod | Multi-provider aggregation | Provider flexibility |
Model Selection Guide
- Best Accuracy
- Cost-Effective
- Speed
- Privacy
For the highest quality scraping results:
- OpenAI GPT-4o: Best overall performance
- Anthropic Claude 3.5 Sonnet: Excellent reasoning
- Google Gemini 2.0 Pro: Great for long documents
Basic Configuration Pattern
All providers follow a similar configuration pattern:The
model field uses the format provider/model-name (e.g., openai/gpt-4o-mini, anthropic/claude-3-5-sonnet-20240620).Provider-Specific Guides
OpenAI
Setup guide for OpenAI models
Ollama
Run models locally
Azure
Azure OpenAI setup
Gemini
Google Gemini configuration
Groq
Groq setup guide
Advanced
Proxy, timeouts, and more
Switching Providers
Switching between providers is simple - just change thellm configuration:
Next Steps
Advanced Configuration
Learn about proxy rotation, custom headers, timeouts, and browser settings
