NVIDIA GPU Issues
CUDA not detected or unavailable
CUDA not detected or unavailable
cudaMallocAsync errors or warnings
cudaMallocAsync errors or warnings
GTX 16 Series (1660, 1650) black outputs or artifacts
GTX 16 Series (1660, 1650) black outputs or artifacts
- GTX 1660 (all variants)
- GTX 1650 (all variants)
- GTX 1630
- T500, T550, T600, T1000, T1200, T2000
- MX450, MX550
- CMP 30HX
RTX 30/40 Series not using Tensor Cores
RTX 30/40 Series not using Tensor Cores
- FP16 accumulation for matrix operations
- cuDNN auto-tuning for optimal kernels
Multiple NVIDIA GPUs - selecting specific GPU
Multiple NVIDIA GPUs - selecting specific GPU
NVIDIA GPU memory leak or not releasing VRAM
NVIDIA GPU memory leak or not releasing VRAM
- Check for circular references in custom nodes (ComfyUI will log warnings about potential memory leaks)
- Manually trigger cleanup via the UI or API
-
Enable deterministic mode to disable some caching:
- Restart ComfyUI to clear all memory
AMD GPU Issues
AMD GPU not detected (Linux)
AMD GPU not detected (Linux)
- Install ROCm following AMD’s official guide
-
Install PyTorch with ROCm support:
-
For latest ROCm 7.2 (may improve performance):
-
Verify detection:
AMD GPU not detected (Windows - RDNA 3/4)
AMD GPU not detected (Windows - RDNA 3/4)
- RDNA 3: RX 7000 series
- RDNA 3.5: Strix Halo, Ryzen AI Max+ 365
- RDNA 4: RX 9000 series
Unsupported AMD GPU architecture override
Unsupported AMD GPU architecture override
AMD GPU slow performance or crashes
AMD GPU slow performance or crashes
-
Enable PyTorch Cross Attention (RDNA3+):
-
Enable TunableOp (first run slow, subsequent faster):
-
For RDNA2 and older, re-enable cuDNN if experiencing issues:
By default, cuDNN is disabled on RDNA3+ for better performance.
-
Check ROCm version compatibility:
- ROCm 7.0+ required for RDNA4
- ROCm 6.4+ recommended for best performance
AMD FP8 support (RDNA4, MI300)
AMD FP8 support (RDNA4, MI300)
- GPU: RDNA4 (RX 9000) or MI300 series
- PyTorch: 2.7+
- ROCm: 6.4+
Intel GPU Issues
Intel Arc GPU not detected
Intel Arc GPU not detected
-
Install Intel Extension for PyTorch:
-
Install PyTorch XPU:
-
For latest features:
-
Verify detection:
Intel Arc performance issues
Intel Arc performance issues
-
IPEX optimization is enabled by default. If experiencing issues:
-
Select specific device:
- Check driver version: Intel frequently updates drivers with performance improvements. Ensure you have the latest Intel graphics drivers.
- FP16 support: ComfyUI automatically detects FP16 capability. On older PyTorch (before 2.3), FP16 is always enabled.
Other Accelerators
Ascend NPU (Huawei)
Ascend NPU (Huawei)
- Ascend Basekit (driver, firmware, CANN)
- torch-npu package
- Install Ascend Basekit following official guide
- Install torch-npu following installation instructions
- Run ComfyUI normally - NPU will be automatically detected
Cambricon MLU
Cambricon MLU
- Cambricon CNToolkit
- torch_mlu package
- Install CNToolkit following official guide
- Install PyTorch MLU following installation guide
- Run ComfyUI - MLU will be automatically detected
Iluvatar Corex
Iluvatar Corex
- Iluvatar Corex Toolkit
- Compatible PyTorch build
- Install Iluvatar Corex Toolkit following official documentation
- Run ComfyUI - Corex will be automatically detected
DirectML (Not Recommended)
DirectML (Not Recommended)
0 with device index. Use -1 for default device.Better alternatives:- Use CPU mode instead:
python main.py --cpu - Use WSL2 with CUDA if on Windows with NVIDIA GPU
- Use native PyTorch CUDA or ROCm
Apple Silicon (MPS)
Apple Silicon GPU issues
Apple Silicon GPU issues
General GPU Diagnostics
Check GPU detection and info
Check GPU detection and info
- Total VRAM
- Total RAM
- Device name and type
- PyTorch version
- CUDA/ROCm/XPU version
- Allocator backend
GPU utilization is low
GPU utilization is low
- CPU bottleneck: Check CPU usage while generating
- Slow storage: Loading models from slow drives
- Wrong precision: Using FP32 when FP16 is supported
- Insufficient VRAM: Models constantly swapping
-
Enable fast mode:
-
Use appropriate VRAM mode:
-
Enable async offloading:
-
Check preview method (previews can slow generation):
Temperature or throttling issues
Temperature or throttling issues
- NVIDIA:
nvidia-smi -l 1 - AMD:
rocm-smi - Intel: Use system monitoring tools
- Improve case airflow
- Clean GPU heatsink and fans
- Reduce power limit if necessary
- Use lower precision to reduce heat: