Overview
OneClaw supports 2 embedding providers for generating dense vector representations of text:| Provider | Default Model | Dimensions | Endpoint |
|---|---|---|---|
| OllamaEmbedding | nomic-embed-text | 768 | http://localhost:11434 |
| OpenAIEmbedding | text-embedding-3-small | 1536 | https://api.openai.com |
OllamaEmbedding
Local, offline-capable embedding generation using Ollama.Configuration
TOML Configuration
Supported Models
| Model | Dimensions | Use Case |
|---|---|---|
nomic-embed-text | 768 | Default - General purpose, good quality |
all-minilm | 384 | Smaller, faster, lower quality |
mxbai-embed-large | 1024 | Higher quality, larger |
snowflake-arctic-embed | 1024 | Alternative high-quality option |
API Format
Endpoint:POST /api/embed
Request:
Usage Example
Performance Characteristics
Latency (localhost):- Single embedding: ~50-200ms
- Batch of 10: ~200-500ms
- RPi 4 (4GB): Works well with
nomic-embed-text(768d) - Desktop/Server: Can use larger models (1024d)
- Fully offline
- No API costs
- Privacy-preserving (data never leaves device)
- Fast for local inference
- Requires Ollama running locally
- Lower quality than OpenAI models
- Requires model to be pulled first (
ollama pull nomic-embed-text)
OpenAIEmbedding
Cloud-based, high-quality embedding generation using OpenAI’s API.Configuration
TOML Configuration
Supported Models
| Model | Dimensions | Cost (per 1M tokens) | Use Case |
|---|---|---|---|
text-embedding-3-small | 1536 | $0.02 | Default - Balanced quality/cost |
text-embedding-3-large | 3072 | $0.13 | Highest quality |
text-embedding-ada-002 | 1536 | $0.10 | Legacy model |
API Key Resolution
Priority order:config.api_key(explicit in code/TOML)OPENAI_API_KEYenvironment variable- Error if none found
API Format
Endpoint:POST /v1/embeddings
Request:
Usage Example
Performance Characteristics
Latency (internet):- Single embedding: ~100-500ms
- Batch of 10: ~200-800ms
- Highest quality embeddings
- No local infrastructure required
- Efficient batch processing
- Reliable uptime
- Requires internet connection
- API costs per token
- Data sent to cloud
- Requires API key management
Comparing Providers
Quality
OpenAI (1536d) > Ollama Large (1024d) > Ollama Default (768d) > Ollama Small (384d)Cost
Ollama (free, local compute) vs OpenAI ($0.02 per 1M tokens)Privacy
Ollama (fully local) vs OpenAI (cloud-based)Use Cases
Choose OllamaEmbedding when:
- Running on edge/IoT devices
- Privacy is critical
- No internet available
- Cost must be zero
- Lower quality is acceptable
Choose OpenAIEmbedding when:
- Highest quality needed
- Internet available
- Budget allows API costs
- Reliability critical