Prerequisites
Before installing Whisper, ensure your system meets these requirements:- Python: Version 3.8 or higher (3.8-3.13 supported)
- PyTorch: Recent versions (developed with PyTorch 1.10.1+)
- FFmpeg: Required for audio processing
Installation Steps
Install Whisper via pip
The easiest way to install Whisper is via pip. This will install the latest stable release:
Alternative: Install from GitHub
To get the latest development version directly from the repository:Install FFmpeg
FFmpeg is required for audio file processing. Install it using your system’s package manager:
Make sure FFmpeg is accessible from your command line by running
ffmpeg -version after installation.Dependencies
Whisper automatically installs the following Python dependencies:- more-itertools: Utilities for working with iterables
- numba: JIT compiler for numerical functions
- numpy: Numerical computing library
- tiktoken: Fast tokenizer implementation by OpenAI
- torch: PyTorch deep learning framework
- tqdm: Progress bar library
- triton: GPU programming (Linux x86_64 only)
Troubleshooting
Rust installation errors
Rust installation errors
If you see installation errors related to
tiktoken, you may need to install Rust:- Visit the Rust installation page
- Follow the installation instructions for your platform
- Configure your PATH:
- Retry the Whisper installation
Missing setuptools_rust
Missing setuptools_rust
If installation fails with Then retry installing Whisper.
No module named 'setuptools_rust', install it explicitly:FFmpeg not found
FFmpeg not found
If you get an error that FFmpeg is not found:
- Verify FFmpeg is installed:
ffmpeg -version - Ensure FFmpeg is in your system PATH
- On Windows, you may need to restart your terminal or add FFmpeg to PATH manually
CUDA/GPU issues
CUDA/GPU issues
Whisper will automatically use CUDA if available. To verify GPU support:If CUDA is not detected but you have an NVIDIA GPU, reinstall PyTorch with CUDA support following the PyTorch installation guide.
SHA256 checksum errors
SHA256 checksum errors
If you encounter SHA256 checksum errors when downloading models:
- Delete the cached model files in
~/.cache/whisper - Retry loading the model
- Check your internet connection for stability
System Requirements by Model
Ensure your system has sufficient VRAM for your chosen model:| Model | VRAM Required | Best For |
|---|---|---|
| tiny | ~1 GB | Fast transcription, limited resources |
| base | ~1 GB | Balanced speed and accuracy |
| small | ~2 GB | Better accuracy, moderate speed |
| medium | ~5 GB | High accuracy, translation tasks |
| large | ~10 GB | Maximum accuracy |
| turbo | ~6 GB | Fast with high accuracy (recommended) |
Next Steps
Quickstart Guide
Start using Whisper with practical examples for both CLI and Python API