Skip to main content

Overview

hw-detect inspects your machine and reports every hardware detail that affects local LLM inference: CPU capabilities, GPU topology (dedicated and integrated), unified or discrete memory, memory bandwidth, and the best available inference backend.
llm-checker hw-detect

Example Output

Apple M4 Pro (24 GB Unified Memory):
=== Hardware Detection ===

Summary:
  Apple M4 Pro (24GB Unified Memory)
  Tier: MEDIUM HIGH
  Max model size: 15GB
  Best backend: metal
  Dedicated GPUs: None
  Integrated GPUs: None

CPU:
  Apple M4 Pro
  Cores: 12 (12 physical)
  SIMD: NEON
  [OK] ARM NEON

METAL:
  Apple M4 Pro
  GPU Cores: 16
  Unified Memory: 24GB
  Memory Bandwidth: 273GB/s

Fingerprint: <hardware-fingerprint>
Hybrid GPU system (dedicated + integrated):
Summary:
  Dedicated GPUs: NVIDIA GeForce RTX 4060
  Integrated GPUs: Intel Iris Xe Graphics
  Assist path: Integrated/shared-memory GPU detected, runtime remains CPU
On hybrid or integrated-only systems, hw-detect surfaces the full GPU inventory. Integrated GPUs are visible even when the selected runtime backend resolves to CPU.

Flags

-j, --json
flag
Output the full hardware detection result as a machine-readable JSON object. Useful for scripting or CI pipelines.

What It Detects

FieldDescription
CPUBrand, core count (logical and physical), clock speed, SIMD capabilities (AVX-512, AVX2, ARM NEON)
GPUModel name, VRAM, GPU type (dedicated vs. integrated), backend availability
MemoryTotal system RAM or unified memory (Apple Silicon)
BackendBest available inference backend: metal, cuda, rocm, cpu
TierHardware tier label: ULTRA LOW, LOW, MEDIUM, HIGH, VERY HIGH, EXTREME
Max model sizeEstimated maximum model size (GB) that fits in available memory

GPU Topology

hw-detect enumerates GPUs into two separate lists:
  • Dedicated GPUs — discrete NVIDIA, AMD, or Intel Arc GPUs with their own VRAM
  • Integrated GPUs — Intel Iris Xe, AMD integrated Radeon, Apple Vega, etc.
On Intel hybrid laptops or mini PCs with both a dedicated GPU and integrated graphics, both appear in the inventory simultaneously. The Assist path message explains when an integrated GPU is visible but the runtime backend still falls through to CPU.
The hardware fingerprint shown at the bottom of the output is a deterministic string that uniquely identifies your hardware configuration. It is used internally for cache keying.

Supported Backends

BackendHardware
metalApple Silicon (M1/M2/M3/M4 series)
cudaNVIDIA GPUs with CUDA support
rocmAMD GPUs with ROCm support
genericIntel Arc and other discrete GPUs
cpuCPU-only inference (fallback)

Usage in Workflows

# Detect hardware and save as JSON
llm-checker hw-detect --json > hardware.json

# Quick human-readable check before running models
llm-checker hw-detect

Build docs developers (and LLMs) love