Supported Providers
OpenAI
GPT-4o and GPT-5 models with vision and extended thinking support
Anthropic
Claude models with strong multilingual capabilities
Gemini models with competitive pricing and performance
Ollama
Local models with zero API costs and full privacy
Model Types
Fromtypes.py:21-28:
OpenAI
OpenAI provides state-of-the-art language models with strong multilingual support and vision capabilities.Available Models
- GPT-5 (2025-08-07): Latest flagship model with extended thinking
- GPT-4o: Fast and cost-effective with vision support
- GPT-4-turbo: Previous generation turbo model
Pricing
Fromcost.py:22-27:
CLI Usage
Environment Setup
Features
- ✅ Vision support (for PDF images)
- ✅ Extended thinking (reasoning models)
- ✅ Checkpoint/resume
- ✅ Glossary support
- ✅ Fast inference
- ✅ High quality output
Anthropic
Anthropic’s Claude models excel at nuanced language understanding and long-form content translation.Available Models
- Claude Sonnet 4: Latest model with extended thinking
- Claude 3.5 Sonnet: Fast and cost-effective
- Claude 3 Opus: Highest quality (previous generation)
Pricing
Fromcost.py:28-31:
CLI Usage
Environment Setup
Claude models are particularly strong at maintaining consistent tone and style across long documents. Ideal for literary translations.
Features
- ✅ Vision support (for PDF images)
- ✅ Extended thinking (reasoning models)
- ✅ Checkpoint/resume
- ✅ Glossary support
- ✅ Strong multilingual support
- ✅ Excellent at literary translation
Google (Gemini)
Google’s Gemini models offer competitive performance with cost-effective pricing.Available Models
- Gemini 2.5 Pro: Latest model with strong multilingual support
- Gemini 1.5 Pro: Previous generation
- Gemini Flash: Fast and economical
Pricing
Fromcost.py:32-35:
CLI Usage
Environment Setup
Gemini models offer the same pricing as OpenAI but with different strengths. Test both to see which works better for your use case.
Features
- ✅ Vision support (for PDF images)
- ✅ Checkpoint/resume
- ✅ Glossary support
- ✅ Fast inference
- ✅ Cost-effective pricing
Ollama (Local Models)
Ollama runs models locally on your machine, providing zero-cost translation with complete privacy.Available Models
Any model available in Ollama can be used:- Llama 3: Strong general-purpose model
- Mistral: Fast and efficient
- Gemma: Google’s open model
- Qwen: Excellent multilingual support
Pricing
Fromcost.py:36:
Setup
- Install Ollama: https://ollama.ai/download
- Pull a model:
- Translate:
CLI Usage
Features
- ✅ Zero cost
- ✅ Complete privacy
- ✅ Checkpoint/resume
- ✅ Glossary support
- ✅ Offline operation
- ❌ No vision support
- ❌ No extended thinking
- ⚠️ Slower than cloud models
- ⚠️ Quality varies by model
Performance Considerations
Fromcost.py:198-200:
Model Selection Syntax
Tinbox supports flexible model selection:Reasoning Effort Levels
Extended thinking models (OpenAI GPT-5, Anthropic Sonnet 4) support reasoning effort configuration: Fromtypes.py:110-113:
Effort Levels
- minimal (default): Standard translation, no extended thinking
- low: Brief reasoning, 2-3x cost
- medium: Moderate reasoning, 5-10x cost
- high: Deep reasoning, 10-20x cost
Cost Tracking
All models (except Ollama) track real-time costs during translation. See Cost Tracking for details.Choosing the Right Provider
When to use OpenAI
When to use OpenAI
- You need vision support for PDF translation
- You want the highest quality output
- You need extended thinking for critical translations
- Cost is not the primary concern
When to use Anthropic
When to use Anthropic
- You’re translating literary or creative content
- You need consistent tone and style
- You value nuanced language understanding
- You want extended thinking capabilities
When to use Google (Gemini)
When to use Google (Gemini)
- You want competitive quality at lower cost
- You need fast inference
- You’re translating general-purpose content
- You want an alternative to OpenAI
When to use Ollama
When to use Ollama
- Cost is a critical factor
- You need complete privacy
- You’re translating large volumes
- You have local compute resources
- You’re willing to accept lower quality
Provider Comparison
| Feature | OpenAI | Anthropic | Ollama | |
|---|---|---|---|---|
| Cost (1K tokens) | 0.01 | 0.015 | 0.01 | Free |
| Vision | ✅ Yes | ✅ Yes | ✅ Yes | ❌ No |
| Reasoning | ✅ Yes | ✅ Yes | ❌ No | ❌ No |
| Speed | Fast | Fast | Fast | Slow |
| Quality | Excellent | Excellent | Very Good | Good |
| Privacy | Cloud | Cloud | Cloud | Local |
| Offline | ❌ No | ❌ No | ❌ No | ✅ Yes |