Cloud Embedding Providers
OpenAI
text-embedding-3-small, text-embedding-3-large, ada-002
Azure OpenAI
OpenAI embeddings via Azure
Cohere
embed-english, embed-multilingual models
Google Gemini
Gemini embedding models
Google Vertex AI
Enterprise embedding models
AWS Bedrock
Titan and Cohere embeddings
Voyage AI
Specialized domain embeddings
Jina AI
Multilingual embedding models
Mistral AI
mistral-embed model
Self-Hosted & Local Embeddings
Ollama
Run embedding models locally
HuggingFace Inference
Access thousands of embedding models
LocalAI
Self-hosted embedding API
Together AI
Open-source embedding models
IBM Watsonx
Enterprise embedding models
Configuration Examples
OpenAI Embeddings
text-embedding-3-small(1536 dims) - Best valuetext-embedding-3-large(3072 dims) - Highest qualitytext-embedding-ada-002(1536 dims) - Legacy
- Get API key from platform.openai.com
- Add credential in Flowise:
- Credential Type:
OpenAI API - API Key:
sk-...
- Credential Type:
Azure OpenAI Embeddings
- Azure OpenAI API Key
- Azure OpenAI API Instance Name
- Azure OpenAI API Deployment Name
- Azure OpenAI API Version
Cohere Embeddings
embed-english-v3.0(1024 dims)embed-multilingual-v3.0(1024 dims)embed-english-light-v3.0(384 dims)
search_query- For queriessearch_document- For documents being indexedclassification- For classification tasksclustering- For clustering tasks
Ollama Embeddings (Local)
-
Install Ollama:
-
Pull embedding model:
-
Available models:
nomic-embed-text(768 dims) - Recommendedmxbai-embed-large(1024 dims)all-minilm(384 dims)
-
Test locally:
HuggingFace Embeddings
sentence-transformers/all-MiniLM-L6-v2(384 dims)sentence-transformers/all-mpnet-base-v2(768 dims)BAAI/bge-small-en-v1.5(384 dims)BAAI/bge-large-en-v1.5(1024 dims)
Google Gemini Embeddings
Voyage AI Embeddings
voyage-2(1024 dims) - General purposevoyage-code-2(1536 dims) - Code embeddingsvoyage-large-2(1536 dims) - Highest quality
Mistral Embeddings
Together AI Embeddings
Advanced Configuration
Custom Dimensions
OpenAI’s new models support custom dimensions:- Smaller storage requirements
- Faster similarity search
- Minimal quality loss
Batch Processing
Process multiple texts efficiently:Custom Base URLs
Use custom endpoints:Timeout Configuration
Embedding Dimensions Guide
| Provider | Model | Dimensions | Use Case |
|---|---|---|---|
| OpenAI | text-embedding-3-small | 1536 | Balanced |
| OpenAI | text-embedding-3-large | 3072 | Best quality |
| Cohere | embed-english-v3.0 | 1024 | General |
| Ollama | nomic-embed-text | 768 | Local |
| HuggingFace | all-MiniLM-L6-v2 | 384 | Fast |
| Voyage | voyage-2 | 1024 | Specialized |
| Mistral | mistral-embed | 1024 | Multilingual |
Choosing an Embedding Model
Best for Quality
- OpenAI text-embedding-3-large - Highest MTEB score
- Voyage voyage-large-2 - Domain-specialized
- Cohere embed-english-v3.0 - Strong retrieval
Best for Speed
- OpenAI text-embedding-3-small - Fast and cheap
- Ollama nomic-embed-text - Local, no latency
- HuggingFace all-MiniLM-L6-v2 - Small model
Best for Cost
- Ollama - Free, run locally
- HuggingFace - Free inference API
- OpenAI text-embedding-3-small - $0.02 per 1M tokens
Best for Privacy
- Ollama - 100% local
- LocalAI - Self-hosted
- HuggingFace (self-hosted) - Deploy yourself
Best for Multilingual
- Cohere embed-multilingual-v3.0 - 100+ languages
- Mistral mistral-embed - Multilingual support
- HuggingFace paraphrase-multilingual - 50+ languages
Performance Optimization
Batch Size Optimization
Caching Embeddings
Cache frequently used embeddings:Dimensionality Reduction
- Reducing storage by 67%
- Speeding up search by 3x
- Lowering costs
Code Examples
Basic Usage
With Vector Store
Custom Embedding Function
Troubleshooting
Dimension Mismatch
Rate Limits
Connection Issues
Out of Memory
Best Practices
- Match dimensions - Embedding and vector store must match
- Batch processing - Use appropriate batch sizes
- Cache embeddings - Avoid re-computing same texts
- Choose right model - Balance quality vs. cost vs. speed
- Test locally first - Use Ollama for development
- Monitor costs - Track API usage and optimize
Next Steps
Vector Stores
Store embeddings in vector databases
Document Loaders
Load documents to embed
