How It Works
World Monitor uses a 4-tier provider fallback chain to ensure you always get a summary, even when cloud services are unavailable:Tier 1: Local LLM (Ollama/LM Studio)
If you’ve configured a local inference endpoint, World Monitor attempts generation first on your machine. This keeps your data private and eliminates API costs.Timeout: 5 secondsModels supported: Any OpenAI-compatible endpoint (Ollama, LM Studio, llama.cpp server, vLLM)
Tier 2: Groq (Cloud)
Fast cloud inference using Llama 3.1 8B at temperature 0.3. Requires a Groq API key.Timeout: 5 seconds
Tier 3: OpenRouter (Cloud)
Multi-model fallback for additional redundancy. Requires an OpenRouter API key.Timeout: 5 seconds
Deduplication & Caching
To optimize performance and reduce API costs:- Content deduplication — Headlines are compared using word-overlap similarity (Jaccard). Near-duplicates (>60% overlap) are merged before sending to the LLM, reducing prompt size by 20–40%
- Redis caching — Summaries are cached for 24 hours with a composite key based on mode, variant, language, and content hash. If 1,000 users view the same headlines, only one LLM call is made
- Variant-aware prompts — The system prompt adapts to your dashboard variant:
- World Monitor: emphasizes geopolitical events, conflict escalation, diplomatic shifts
- Tech Monitor: focuses on funding rounds, AI breakthroughs, startup news
- Finance Monitor: highlights market movements, central bank signals, economic data
Multilingual Output
When you change the UI language in World Monitor, the World Brief automatically generates summaries in that language. The LLM prompt instructs the model to output in your selected language:- English
- French
- Arabic
Configuring Local LLM
To use Ollama or LM Studio for local generation:- Open Settings (Cmd+, or click the gear icon)
- Navigate to the LLMs tab
- Enter your local endpoint URL:
- Ollama default:
http://localhost:11434 - LM Studio default:
http://localhost:1234
- Ollama default:
- Click Verify & Save
Progress Indicators
While generating a summary, World Monitor displays which provider tier is being attempted:- “Generating summary (Ollama)…” — Trying local endpoint
- “Generating summary (Groq)…” — Trying cloud inference
- “Generating summary (OpenRouter)…” — Trying multi-model fallback
- “Generating summary (Browser)…” — Using local browser model
Privacy Considerations
- Ollama/LM Studio (Tier 1)
- Groq/OpenRouter (Tiers 2-3)
- Browser T5 (Tier 4)
✅ 100% private — No data leaves your machine
✅ No API keys required
✅ No internet required (after model download)
✅ No API keys required
✅ No internet required (after model download)
API Key Management
API keys for cloud providers are stored securely:- Desktop app: OS keychain (macOS Keychain, Windows Credential Manager)
- Web app: Browser localStorage (encrypted in transit via HTTPS)
Troubleshooting
Summary always falls back to browser model
Summary always falls back to browser model
This means all three API tiers are timing out or returning errors:
- Check your Settings → LLMs tab to verify Ollama/Groq/OpenRouter are configured
- Test your Ollama endpoint:
curl http://localhost:11434/api/tags - Verify API keys are valid (click Verify & Save in Settings)
- Check browser console for detailed error messages
Ollama tier always skipped
Ollama tier always skipped
Ensure:
- Ollama is running:
ollama serve - The endpoint URL is correct in Settings (default:
http://localhost:11434) - At least one chat model is installed:
ollama list - The model isn’t embedding-only (e.g.,
nomic-embed-textwon’t work)
Summaries are in the wrong language
Summaries are in the wrong language
The UI language setting controls summary language. Change it via the language selector in the header:
- Click the language flag icon (e.g., 🇬🇧)
- Select your preferred language
- The page will reload with the new language
- Generate a new World Brief — it will now use your language
Summary is stale or outdated
Summary is stale or outdated
Summaries are cached for 24 hours. To force regeneration:
- Wait for new headlines to arrive (cache key includes content hash)
- Or clear your browser cache and reload
Related Features
- Country Analysis — AI-generated briefs for specific countries
- Signal Intelligence — Real-time threat detection and convergence alerts
- Localization — Changing UI language and regional settings