Top-Level Commands
serve
Load a local model and expose it at localhost:6767/v1. Auto-detects LlamaCPP or MLX.launch
Start a local model, then launch an AI agent with it pre-wired (environment variables set automatically).threads
List and inspect conversation threads saved by the Jan app.Subcommands
threads list
threads list
Print all threads as JSON.Output: JSON array of thread objects with metadata.
threads get <ID>
threads get <ID>
Print a single thread’s metadata as JSON.Output: JSON object with thread metadata.
Thread ID to retrieve
threads delete <ID>
threads delete <ID>
Permanently delete a thread and all its messages.Output: Confirmation JSON with deleted thread ID.
Thread ID to delete
threads messages <ID>
threads messages <ID>
Print all messages in a thread as JSON.Output: JSON array of message objects.
Thread ID to get messages from
models
List and load models installed in the Jan data folder.Subcommands
models list
models list
Print all installed models as JSON (from the Jan data folder).Example Output:
Filter by engine:
llamacpp, mlx, or allmodels load
models load
Load a model and serve it — alias for the top-level Accepts the same options as
serve command.jan serve. See serve documentation.models load-mlx
models load-mlx
Load an MLX model directly (macOS / Apple Silicon only).
Model ID as shown by
jan models list --engine mlxPath to the MLX model directory (auto-resolved from model.yml when omitted)
Path to the mlx-server binary (auto-discovered from Jan.app when omitted)
Port the model server listens on (0 = pick a random free port)
Context window size in tokens (0 = model default)
Treat the model as an embedding model
Seconds to wait for the model server to become ready
API key required by clients (sets MLX_API_KEY on the server)
app
Show app configuration and data folder location.Subcommands
app data-folder
app data-folder
Print the Jan data folder path (where models, threads, and config are stored).Example Output:
app config
app config
Print the Jan configuration as JSON.Output: JSON object with Jan configuration settings.
Global Options
Display help information for any command
Display the Jan CLI version
Print full server logs (llama.cpp / mlx output) instead of the loading spinnerAvailable on:
serve, launchExamples
Exit Codes
0: Success1: Error (with error message printed to stderr)
See Also
Serve Command
Detailed guide for serving models
Launch Command
Wire AI agents to local models