Overview
recommend generates hardware-aware model recommendations across multiple categories simultaneously: Coding, Reasoning, Multimodal, Creative, Chat, and more. It uses the deterministic 4D scoring engine and optionally routes through a calibrated policy for measured routing decisions.
Example Output
Flags
Narrow output to a single category. See the category table below for accepted values.
Apply an optimization profile to steer ranking. Accepted values:
balanced, speed, quality, context, coding.Default: balancedEnable calibrated routing. Optionally provide a file path to a calibration policy. If omitted, discovery checks
~/.llm-checker/calibration-policy.{yaml,yml,json}.Enterprise policy file path. Takes precedence over
--calibrated.Disable step-by-step progress output.
Simulate a hardware profile. Use
--simulate list to see all profiles.Custom GPU model for hardware simulation, e.g.
"RTX 5060".Custom RAM in GB for hardware simulation.
Custom CPU model for hardware simulation.
Override GPU VRAM in GB. Requires
--gpu.Category Options
| Category | Use Case |
|---|---|
coding | Programming, code generation, code review |
reasoning | Complex logic, math, multi-step problems |
multimodal | Image understanding, vision tasks |
creative | Creative writing, storytelling |
chat | Conversational AI, general chat |
general | General-purpose tasks |
embeddings | Semantic search, RAG pipelines |
Optimize Profiles
| Profile | Description |
|---|---|
balanced | Equal emphasis on quality and speed |
speed | Maximize tokens/sec, prefer smaller models |
quality | Maximize model quality, accept slower inference |
context | Prefer models with large context windows |
coding | Emphasize coding benchmark scores |
Usage Examples
Calibrated Routing
When--calibrated is active, routing decisions are sourced from a calibration-policy.yaml generated by the calibrate command. The output includes a CALIBRATED ROUTING block showing:
- Policy file path and discovery source
- Task name (and any task fallback used)
- Route primary model and fallbacks
- Selected model
--policy <file>(explicit enterprise policy)--calibrated <file>(explicit calibration policy file)--calibrated(auto-discovery from~/.llm-checker/)- Deterministic selector fallback
The fine-tuning label in output (
LoRA, QLoRA, Full FT) reflects the suitability of the recommended model for fine-tuning workflows based on parameter count and quantization.
