Syntax
Description
Search HuggingFace for GGUF models that are compatible with llama.cpp. This command queries the HuggingFace Hub API to find repositories containing GGUF files, which are quantized model formats optimized for local inference. Use this command to discover available models before downloading them withllmfit download.
Arguments
Search query. Can include model name, architecture, or other keywords (e.g.,
llama, mistral, coding).Options
Maximum number of results to return. Range: 1-100.
Usage Examples
Search for Llama models
Search for coding models with limit
Search for specific architecture
Search for quantized versions
Example Output
Output Fields
Each result includes:- Repository name: Full HuggingFace repo identifier (user/repo)
- Download count: Total number of downloads from the Hub
- Likes: Number of likes/stars on HuggingFace
- Last updated: Most recent update date
- Tags: Repository tags indicating model type and format
Search Tips
Results are sorted by relevance and popularity. The most popular and recently updated models appear first.
Notes
- Only repositories containing GGUF files are returned
- Results include official quantizations from popular providers like bartowski, TheBloke, QuantFactory, and lmstudio-community
- The search queries the HuggingFace Hub API, so an internet connection is required
- Repository metadata (downloads, likes) is fetched in real-time
Related Commands
- llmfit download - Download a GGUF model from HuggingFace
- llmfit search - Search llmfit’s built-in model database
- llmfit list - List all models in llmfit’s database
