Overview
Off Grid is designed to work completely offline after initial model downloads. This guide explains what works offline, what requires network, and how to optimize for offline use.What Works Offline
FromARCHITECTURE.md:1044-1047, after model download:
✅ All AI inference - Text generation, vision, image generation
✅ Conversations - Create, edit, delete conversations
✅ Projects - Create and manage custom system prompts
✅ Model management - Load, unload, delete models
✅ Settings - Change all app settings
✅ Voice transcription - Whisper runs entirely on-device
✅ Document attachments - Attach and analyze PDFs, code files, etc.
✅ Image generation - Stable Diffusion runs on-device (CPU/GPU/NPU/ANE)
✅ Tool calling - Calculator, date/time, device info tools
✅ Gallery - View, share, delete generated images
Everything works offline except model downloads and the optional web search tool.
What Requires Network
FromARCHITECTURE.md:1036-1042:
During Setup
❌ HuggingFace API - Model metadata browsing ❌ HuggingFace CDN - Model file downloads ❌ xororz HuggingFace repos - Image model listings (Android)Optional Features
❌ Web Search tool - Requires explicit user enablement, uses Brave SearchAirplane Mode Usage
Complete Offline Workflow
Download Models (Requires Network)
While connected to Wi-Fi:
- Browse and download text models (e.g., Qwen 3 0.6B)
- Download vision models if needed (SmolVLM-500M)
- Download image models if needed (Anything V5, DreamShaper)
- Download Whisper model for voice (auto-downloads on first use)
Use Off Grid Normally
Everything works:
- Load models
- Chat conversations
- Generate images
- Transcribe voice
- Attach documents
- Create projects
Storage Planning for Offline Use
Calculate Storage Needs
Before going offline, ensure you have enough storage: Text Models:- Tiny (0.5-1B): 0.5-1 GB
- Small (1-3B): 1-2 GB
- Medium (3-7B): 2-5 GB
- Large (7-8B): 4-8 GB
- SmolVLM-500M: ~600 MB (model + mmproj)
- SmolVLM-2.2B: ~2.5 GB (model + mmproj)
- Qwen3-VL-2B: ~2.8 GB
- Android CPU (MNN): ~1.2 GB per model
- Android NPU (QNN): ~1.0 GB per model
- iOS Palettized: ~1 GB per model
- iOS Full precision: ~4 GB per model
- Tiny: ~75 MB
- Base: ~140 MB
- Small: ~470 MB
Storage Planning Examples
Minimal Offline Setup (2-3 GB):- 1 text model (Qwen 3 0.6B Q4_K_M): ~500 MB
- Whisper Tiny: ~75 MB
- Buffer for conversations and images: ~500 MB
- 2 text models (Qwen 3 3B Q4_K_M + Phi-4 Mini): ~4 GB
- 1 vision model (SmolVLM-500M): ~600 MB
- 1 image model (Anything V5): ~1.2 GB
- Whisper Base: ~140 MB
- Buffer: ~1 GB
- 4 text models (variety of sizes and specializations): ~10 GB
- 2 vision models (SmolVLM-500M + Qwen3-VL-2B): ~3.4 GB
- 3 image models (variety of styles): ~3.6 GB
- Whisper Small: ~470 MB
- Buffer: ~2 GB
Model Selection for Offline Use
Choosing Models for Limited Storage
FromARCHITECTURE.md:950-961, quantization trade-offs:
| Priority | Model Choice | Quantization | Size (7B) | Quality |
|---|---|---|---|---|
| Maximum quality | Larger models | Q6_K, Q8_0 | 6-7.5 GB | Excellent |
| Balanced | Medium models | Q4_K_M, Q5_K_M | 4-5 GB | Good-Very Good |
| Space-constrained | Smaller models | Q3_K_M, Q4_K_M | 3-4 GB | Low-Med to Good |
| Minimum storage | Tiny models | Q4_K_M | 0.5-2 GB | Good (for size) |
Recommended Offline Model Sets
General Purpose (Balanced):- Qwen 3 3B Q4_K_M (~2.5 GB) - Main chat model
- SmolVLM-500M (~600 MB) - Vision tasks
- Anything V5 (~1.2 GB) - Image generation
- Whisper Base (~140 MB) - Voice transcription
- Qwen 3 Coder A3B Q4_K_M (~2.8 GB)
- Phi-4 Mini Q4_K_M (~2.5 GB)
- Qwen3 7B Q4_K_M (~4.8 GB)
- Command-R 3B Q4_K_M (~2.2 GB)
- Gemma 3 9B Q4_K_M (~5.5 GB)
- Llama 3.2 3B Q5_K_M (~2.8 GB)
Offline-First Features
Voice Transcription
FromARCHITECTURE.md:143-159, Whisper runs entirely on-device:
- Multiple models - Tiny, Base, Small (speed vs accuracy)
- Real-time transcription - Streaming word-by-word results
- No network - All transcription on-device
- Auto-download - First use downloads model (requires network once)
- Multilingual - Supports many languages
- Download Whisper model while online (Settings → Voice Settings)
- Use voice input in airplane mode
- Transcription works perfectly offline
Document Attachments
FromARCHITECTURE.md:160-188, document analysis is fully offline:
- File picker - Native file selection
- PDF text extraction - Platform-native parsers (no network)
- Code files - .py, .js, .ts, .jsx, .tsx, .java, .c, .cpp, etc.
- Data files - .csv, .json, .xml, .yaml, etc.
- Tappable badges - Open with system viewer (QuickLook/Intent)
- Attach document in chat (PDF, code file, etc.)
- AI reads document content
- Ask questions about the document
- All analysis happens on-device
Image Generation
FromARCHITECTURE.md:71-142, Stable Diffusion runs entirely on-device:
Android:
- MNN backend (CPU, all devices)
- QNN backend (NPU, Snapdragon 8 Gen 1+)
- Core ML backend (Neural Engine acceleration)
- Palettized models (6-bit, ~1GB)
- Full precision models (fp16, ~4GB)
- Download image model while online
- Generate images in airplane mode
- All generation on-device (CPU/GPU/NPU/ANE)
Tool Calling (Offline Tools)
FromARCHITECTURE.md:189-218, several tools work offline:
✅ Calculator - Safe recursive descent parser, no network
✅ Date/Time - System date/time, no network
✅ Device Info - Battery, storage, memory stats, no network
❌ Web Search - Requires network (Brave Search API)
Airplane Mode Test
Verification Checklist
Prove to yourself that Off Grid is truly offline:Download Required Models
While online:
- At least one text model
- Optional: vision model, image model, Whisper model
Test Text Generation
- Load text model
- Create conversation
- Send multiple messages
- Verify responses generate normally
Test Vision (If Downloaded)
- Load vision model
- Attach image from gallery
- Ask about the image
- Verify vision inference works
Test Image Generation (If Downloaded)
- Ensure image model is active
- Send image generation prompt
- Verify image generates offline
Test Voice Transcription (If Downloaded)
- Tap voice input button
- Record audio
- Verify transcription works
Test Document Attachment
- Attach a PDF or text file
- Ask questions about it
- Verify AI reads and responds
If any feature fails, ensure:
- Model is fully downloaded (not partial)
- Model is loaded (active in model selector)
- Web search tool is disabled (if testing non-web tools)
Network Monitoring (Advanced)
For technical users who want to verify zero network activity:iOS Network Monitoring
- Use Charles Proxy or Proxyman
- Configure device to use proxy
- Enable airplane mode
- Use Off Grid normally
- Verify zero network requests in proxy logs
Android Network Monitoring
- Use Packet Capture app (no root required)
- Start packet capture
- Enable airplane mode
- Use Off Grid normally
- Verify zero network packets
Alternative: System Network Stats
iOS:- Settings → Cellular → Off Grid
- Note current data usage
- Use app in airplane mode
- Re-check data usage (should be unchanged)
- Settings → Network & Internet → Data Usage → Off Grid
- Note current usage
- Use app in airplane mode
- Re-check usage (should be unchanged)
Privacy Benefits of Offline Use
From the Privacy & Security guide, offline usage provides: ✅ Zero data exfiltration - Impossible when network is disabled ✅ No tracking - No telemetry or analytics can phone home ✅ Guaranteed privacy - Prompts never leave device ✅ Air-gapped security - Perfect for sensitive work ✅ Verifiable privacy - You can see network is offTroubleshooting Offline Issues
Model Won’t Load Offline
Symptoms: Model fails to load in airplane mode Solutions:- Verify model download completed (Settings → Storage Settings)
- Check model file exists on device
- Restart app and try again
- Re-download model while online
Vision Model Fails Without Network
Symptoms: Vision model loads but errors on image analysis Solutions:- Verify mmproj file downloaded (check model size includes both files)
- Reload model to trigger mmproj discovery
- Check logs for mmproj path resolution
Web Search Tool Errors
Symptoms: Tool calling fails with network errors Solutions:- Disable web search tool in settings
- Use only offline tools (calculator, date/time, device info)
- Model will gracefully skip web search when network unavailable
Whisper Transcription Fails
Symptoms: Voice input doesn’t transcribe offline Solutions:- Ensure Whisper model downloaded (Settings → Voice Settings)
- Check model selection (Tiny, Base, or Small)
- Re-download Whisper model while online
Best Practices for Offline Use
Before Going Offline
Download All Required Models
- Text model(s) for your use cases
- Vision model if you need image analysis
- Image model if you want to generate images
- Whisper model for voice input
While Offline
✅ Monitor storage - Generated images consume space ✅ Delete old conversations - Free up space if needed ✅ Use storage cleanup - Remove orphaned files periodically ✅ Keep device charged - Offline AI is power-intensiveGoing Back Online
✅ Download new models if needed ✅ Check for app updates (optional) ✅ Sync device backup (if using iCloud/Google Drive)Additional Resources
- Privacy Guide: See Privacy & Security guide for detailed privacy features
- Model Management: See Model Management guide for download details
- Architecture Reference: Review
ARCHITECTURE.md:1036-1047for network activity details