Hardware Advisor
SlasshyWispr includes an intelligent hardware advisor that analyzes your system and provides model recommendations.What Gets Analyzed
The hardware advisor collects:The hardware advisor runs automatically when you first access local model settings. Results are cached to avoid repeated system scans.
CPU Requirements
Minimum Requirements
- Processor: Modern x86_64 or ARM64 CPU
- Cores: 2 logical cores minimum
- Instruction Sets: AVX support recommended for optimal performance
Recommended CPUs
Entry Level
4-6 coresIntel Core i3/i5, AMD Ryzen 3/5, Apple M1Suitable for: Moonshine, SenseVoice, Whisper Small
Mid Range
6-8 coresIntel Core i5/i7, AMD Ryzen 5/7, Apple M2/M3Suitable for: Parakeet, Whisper Medium, small Ollama models
High End
8+ coresIntel Core i7/i9, AMD Ryzen 7/9, Apple M2/M3 Pro/MaxSuitable for: All models, large Ollama models
CPU Performance Impact
- STT Transcription: CPU-intensive for model inference
- Ollama Inference: Benefits greatly from more cores
- Concurrent Operations: Multiple cores enable simultaneous STT + AI
Apple Silicon Macs (M1/M2/M3) provide excellent performance due to unified memory architecture and Neural Engine acceleration.
RAM Requirements
Memory Guidelines
RAM requirements depend on model sizes:| Total RAM | STT Models | Ollama Models |
|---|---|---|
| 4 GB | Moonshine only | Not recommended |
| 8 GB | Moonshine, SenseVoice, Whisper Small | 1B-3B models |
| 16 GB | All small/medium models, Parakeet | 7B-8B models |
| 32 GB | All models including Whisper Turbo | 13B-14B models |
| 64+ GB | All models with headroom | 30B+ models |
RAM Usage Patterns
STT Model Loading
Models consume RAM approximately equal to their file size:
- Moonshine: ~60 MB
- SenseVoice: ~160 MB
- Parakeet/Whisper Small/Medium: ~500 MB
- Whisper Large/Turbo: 1.1-1.6 GB
Ollama Model Loading
LLM models consume significant RAM:
- 1B-3B models: 2-4 GB
- 7B-8B models: 6-10 GB
- 13B-14B models: 12-18 GB
- 30B+ models: 25+ GB
GPU Support
NVIDIA GPU Detection
SlasshyWispr automatically detects NVIDIA GPUs and reports:- GPU model name
- VRAM capacity
- CUDA availability
GPU Acceleration Benefits
STT Acceleration
NVIDIA GPUs can accelerate ONNX model inferenceSpeed improvement: 2-5x faster transcription
Ollama Acceleration
Ollama automatically uses GPU for LLM inferenceSpeed improvement: 5-20x faster responses
Supported GPUs
NVIDIA (CUDA):- GTX 1000 series and newer
- RTX 2000, 3000, 4000 series
- Professional GPUs (Quadro, Tesla, A-series)
- Requires: CUDA 11.x or newer
- M1, M2, M3 series (integrated GPU)
- Metal acceleration automatic on macOS
- Limited support via Ollama on Linux
- Not supported for ONNX Runtime in SlasshyWispr
VRAM Requirements
| VRAM | STT Models | Ollama Models |
|---|---|---|
| 2-4 GB | Small models | 1B-3B models |
| 6-8 GB | Medium models | 7B-8B models |
| 12 GB | Large models | 13B-14B models |
| 16+ GB | All models | 30B+ models |
If VRAM is insufficient, models will fall back to CPU/RAM. This is slower but still functional.
Performance Tiers
Based on your hardware analysis, SlasshyWispr assigns a performance tier:Tier 1: Entry Level
Hardware:- 4-8 GB RAM
- 2-4 CPU cores
- No dedicated GPU
- STT: Moonshine Base, SenseVoice
- AI: Ollama not recommended, use online mode
Tier 2: Standard
Hardware:- 8-16 GB RAM
- 4-6 CPU cores
- Optional: Entry-level GPU (2-4GB VRAM)
- STT: Parakeet v3, Whisper Small/Medium
- AI: Ollama 1B-3B models (llama3.2:1b, mistral:3b)
Tier 3: Performance
Hardware:- 16-32 GB RAM
- 6-8 CPU cores
- Recommended: Mid-range GPU (6-8GB VRAM)
- STT: Whisper Medium, Parakeet v3, Whisper Turbo
- AI: Ollama 7B-8B models (llama3.2, mistral, gemma2)
Tier 4: Enthusiast
Hardware:- 32+ GB RAM
- 8+ CPU cores
- High-end GPU (12+ GB VRAM)
- STT: Any model, Whisper Turbo for best accuracy
- AI: Ollama 13B-30B+ models (llama3.2:13b, mixtral)
Your performance tier is calculated automatically and displayed in Settings > Offline along with customized model recommendations.
Model Recommendations Based on Hardware
Suggestion Algorithm
SlasshyWispr recommends models based on:- Total RAM: Must fit model + OS + headroom
- CPU cores: More cores = better performance with larger models
- GPU presence: Enables acceleration tier recommendations
- VRAM: Determines max GPU-accelerated model size
Suggested Models Array
ThesuggestedModels array contains models that should work well on your system, ordered by recommendation priority.
Example (16GB RAM, 8 cores, RTX 3060):
Caution Models Array
ThecautionModels array lists models that may struggle on your hardware:
- May exceed available RAM
- Could cause slow performance
- Might require swap/paging
Selected Model Warning
If you choose a model outside recommendations,selectedModelWarning provides specific guidance:
- “This model may exceed available RAM”
- “Consider a smaller model for better performance”
- “GPU acceleration recommended for this model”
- Empty string if selection is optimal
Checking Your Hardware
View Hardware Analysis
Your hardware details are displayed:
- CPU model and core count
- Total RAM
- GPU detection status
- Performance tier
Review Recommendations
SlasshyWispr highlights:
- Top recommended model (green)
- Other suggested models
- Models to avoid (red/yellow warnings)
Optimizing Performance
For Low-Resource Systems
For High-Performance Systems
Storage Requirements
Disk Space Needed
- Application: ~300 MB
- STT Models: 58 MB - 1.6 GB per model
- Ollama Models: 1 GB - 40+ GB per model
- Voice Models (Piper): ~10-50 MB per voice
- Cache & Logs: ~100-500 MB
Use SSD storage for best model loading performance. HDDs will work but may increase warmup times.
Troubleshooting Hardware Issues
GPU Not Detected
Issue: NVIDIA GPU not showing in hardware advisor Solutions:- Install latest NVIDIA drivers
- Install CUDA toolkit (11.x or 12.x)
- Restart SlasshyWispr after driver installation
- Check GPU visibility:
nvidia-smi(Linux/Windows)
Insufficient RAM Warnings
Issue: Cannot load model, out of memory errors Solutions:- Close other applications
- Choose a smaller model
- Increase system swap space (temporary solution)
- Upgrade RAM for better experience
Slow Performance
Issue: Models load but transcription/inference is slow Solutions:- Check if swap/paging is active (indicates RAM pressure)
- Reduce model size
- Enable GPU acceleration if available
- Close background processes
- Check CPU throttling (thermal issues)
Model Warmup Failures
Issue: Model fails to warm up or load Solutions:- Verify sufficient disk space for model files
- Check available RAM exceeds model size + 2GB
- Review logs for specific errors
- Try re-downloading the model
- Update to latest SlasshyWispr version
The hardware advisor automatically refreshes recommendations when you install new RAM, upgrade GPU, or change system configuration.