Quickstart
Get up and running with Jan in minutes. This guide takes you from installation to your first AI conversation with a local model.Install Jan
Download Jan for your platform and install it.Alternative Download Options:
- Tab Title
- Tab Title
- Tab Title
- Download jan.exe
- Run the installer
- Wait for installation to complete
- Launch Jan
Windows 10 or higher required. GPU acceleration available for NVIDIA, AMD, and Intel Arc GPUs.
- Official website: jan.ai/download
- GitHub releases: github.com/janhq/jan/releases
- Microsoft Store (Windows)
- Flathub (Linux)
Download Your First Model
We recommend starting with Jan v1, a 4B parameter model optimized for reasoning and tool calling.Other Popular Models:
- Open Jan
- Navigate to the Hub tab in the left sidebar
- Search for “Jan v1”
- Choose a quantization based on your available RAM:
- Q4_K_M (2.5 GB) - Good balance for most users
- Q8_0 (4.28 GB) - Best quality if you have 16GB+ RAM
- Click Download
- Wait for the download to complete
Jan v1 achieves 91.1% accuracy on SimpleQA and excels at tool calling, making it perfect for web search and reasoning tasks.
- Llama 3.2 3B - Fast, general-purpose model (2GB)
- Gemma 2 2B - Google’s efficient model (1.6GB)
- Qwen 2.5 7B - Excellent for coding and reasoning (4.5GB)
- DeepSeek R1 - Advanced reasoning model (8GB+)
What’s Next?
You’re now chatting with AI completely offline! Here are some powerful features to explore:Enable Web Search (MCP)
Give Jan v1 access to current web information with real-time search:- Go to Settings → MCP
- Add the Serper MCP server (includes 2,500 free searches)
- Enable tools in your chat
- Ask Jan to search the web for current information
Serper MCP Tutorial
Complete guide to enabling web search in Jan with 2,500 free searches per month.
Manage Your Conversations
Jan organizes all your chats into threads for easy management: View Chat History:- Left sidebar shows all conversations
- Favorites: Pin important threads by clicking the star icon
- Recents: Quick access to recently used threads
- Hover over a conversation in the sidebar
- Click the three dots (…) menu
- Select “Rename”
- Enter a descriptive title
- Hover over a thread
- Click three dots (…)
- Select “Delete”
Create Custom Assistants
Customize how models respond to match your workflow:- Click the Assistant tab in the left sidebar
- Click Create New Assistant
- Give it a name (e.g., “Code Reviewer”, “Writing Assistant”)
- Add custom instructions:
- Save and select your assistant in any chat
Assistant instructions work across all models, so you can switch models while keeping the same personality and behavior.
Adjust Model Parameters
Fine-tune model behavior for different tasks:- In any chat, click the Gear icon next to the model name
- Adjust these key parameters:
- Temperature (0.0-1.0): Lower = more focused, higher = more creative
- Max Tokens: Maximum length of responses
- Top P: Controls diversity of word selection
- Frequency Penalty: Reduces repetition
- Creative Writing: Temperature 0.8-1.0
- Code Generation: Temperature 0.2-0.4
- Factual Q&A: Temperature 0.1-0.3
- General Chat: Temperature 0.7 (default)
Download More Models
Explore thousands of models from HuggingFace:- Go to the Hub tab
- Browse by category or search for specific models
- Popular categories:
- General Chat: Llama, Gemma, Mistral
- Coding: Qwen Coder, DeepSeek Coder
- Reasoning: DeepSeek R1, Jan v1
- Small & Fast: Phi-3, TinyLlama
Memory Requirements:
- 8GB RAM → up to 3B parameter models
- 16GB RAM → up to 7B parameter models
- 32GB RAM → up to 13B parameter models
Connect Cloud Models (Optional)
Combine local and cloud AI in one interface:- Open any chat
- Click the model dropdown
- Select a cloud provider (OpenAI, Anthropic, Groq, Mistral)
- Click the Gear icon next to the provider name
- Enter your API key
- Start chatting with GPT-4, Claude, or other cloud models
- OpenAI: GPT-4, GPT-4 Turbo, GPT-3.5
- Anthropic: Claude 3.5 Sonnet, Claude 3 Opus/Haiku
- Groq: Ultra-fast inference with Llama, Mixtral
- Mistral AI: Mistral Large, Medium, Small
- Cohere: Command R+, Command R
- Google: Gemini Pro, Gemini Flash
Remote Models Guide
Detailed setup instructions for connecting cloud AI providers.
Use the Local API Server
Jan runs an OpenAI-compatible API server atlocalhost:1337:
- Continue.dev (VS Code AI coding assistant)
- Tabby (AI code completion)
- n8n (workflow automation)
- LLMCord (Discord bot)
- Custom applications
API Server Examples
See real-world examples of integrating Jan’s API with popular tools.
Troubleshooting
Model Won’t Load
- Check RAM: Ensure you have enough memory for the model size
- GPU Settings: Go to Settings → Hardware → GPUs to enable/disable GPU acceleration
- Restart Jan: Sometimes a fresh start helps
Slow Performance
- Enable GPU acceleration (Settings → Hardware)
- Use smaller models (3B-7B instead of 13B+)
- Lower quantization (Q4 instead of Q8)
- Close other applications to free up RAM
Download Issues
- Check internet connection
- Verify disk space (10GB+ free recommended)
- Try different mirror: Some models have multiple download sources
- Resume failed downloads: Jan automatically resumes interrupted downloads
GPU Not Detected
- Tab Title
- Tab Title
- Tab Title
- Install NVIDIA drivers (version 470.63.01+)
- Install CUDA Toolkit (version 11.7+)
- Verify installation:
- Enable in Jan: Settings → Hardware → GPUs → Toggle ON
Full Troubleshooting Guide
Comprehensive troubleshooting documentation for common issues.
Getting Help
If you run into issues:- Check the troubleshooting docs
- Search GitHub issues
- Ask in Discord #🆘|jan-help channel
- Include:
- Your OS and version
- Jan version (Help → About)
- Error messages or logs
- Steps to reproduce
Next Steps
Installation Guide
Platform-specific installation details and GPU setup instructions.
Model Parameters
Learn how to fine-tune model behavior for different use cases.
MCP Integration
Enable agentic capabilities with web search, tools, and integrations.
API Reference
Complete API documentation for building integrations.