HuggingFaceEmbeddings class provides integration with HuggingFace’s sentence-transformers library for local embedding generation.
Installation
Usage
Basic usage
Embed single text
Embed multiple texts
Configuration
Supported models
Any sentence-transformers model from HuggingFace Hub:sentence-transformers/all-mpnet-base-v2- High quality, 768 dimensionssentence-transformers/all-MiniLM-L6-v2- Fast and efficient, 384 dimensionssentence-transformers/paraphrase-multilingual-MiniLM-L12-v2- Multilingual supportBAAI/bge-large-en-v1.5- State-of-the-art English embeddingsBAAI/bge-small-en-v1.5- Smaller, faster alternative
Device selection
Run embeddings on GPU:Advanced options
Multi-GPU processing
Enable multi-GPU processing:Custom cache folder
Parameters
Name of the sentence-transformers model to use.
Path to store models. Can also be set by
SENTENCE_TRANSFORMERS_HOME environment variable.Keyword arguments for the SentenceTransformer model, such as
device, trust_remote_code, or token.Keyword arguments for the
encode method, such as batch_size, normalize_embeddings, or precision.Separate keyword arguments for query encoding. Falls back to
encode_kwargs if not provided.Run encoding on multiple GPUs.
Whether to show a progress bar during embedding.