Overview
VecLabs delivers sub-5ms p99 query latency at scale with 88% lower costs than Pinecone. Our Rust HNSW implementation has no garbage collector, eliminating latency spikes under concurrent load that plague Python and Go-based vector databases.All benchmarks measured on Apple M2, 16GB RAM. Real hardware, real measurements.
Query Performance
Standard Benchmark (100K vectors, 384 dimensions, top-10)
VecLabs
- p50: 1.9ms
- p95: 2.8ms
- p99: 4.3ms
- Language: Rust (no GC)
Pinecone s1
- p50: ~8ms
- p95: ~15ms
- p99: ~25ms
- Language: Go/Python
Qdrant
- p50: ~4ms
- p95: ~9ms
- p99: ~15ms
- Language: Rust
Weaviate
- p50: ~12ms
- p95: ~25ms
- p99: ~40ms
- Language: Go
Why VecLabs Is Faster
No Garbage Collection
No Garbage Collection
Python and Go have garbage collectors that introduce unpredictable latency spikes during query execution. Rust has no GC—memory is managed at compile time. This means consistent, predictable sub-millisecond latency even under high concurrent load.
Pure HNSW in Rust
Pure HNSW in Rust
Our HNSW (Hierarchical Navigable Small World) graph implementation is built from scratch in Rust with zero runtime overhead. No abstraction layers, no VM, no interpreter—just native machine code executing graph traversal at CPU speed.
In-Memory Query Layer
In-Memory Query Layer
The query engine runs entirely in memory with serialization support for persistence. No network hops to remote storage during query execution. Vector retrieval happens at RAM speed.
Benchmark by Dimension
Query performance across different embedding dimensions (10K vectors, top-10 results):| Dimensions | Use Case | p50 Latency | p99 Latency |
|---|---|---|---|
| 128 | Lightweight embeddings | < 1ms | < 2ms |
| 384 | MiniLM, all-MiniLM-L6-v2 | 1.9ms | 4.3ms |
| 768 | BERT-base, Sentence-BERT | 2.1ms | 4.8ms |
| 1536 | OpenAI text-embedding-3-small | 2.4ms | 5.2ms |
VecLabs performance scales linearly with dimension size. Higher dimensional embeddings require more distance calculations but remain under 5ms p99.
Cost Comparison
Monthly Cost: 1 Million Vectors
VecLabs
~$8/month
- Storage: ~$0.04 (Shadow Drive)
- Merkle updates: ~$0.00025/tx
- Query compute: User infrastructure
- 88% cheaper than Pinecone
Pinecone s1
$70/month
- Storage: Included
- API access: Included
- Query compute: Pinecone cloud
- Lock-in to proprietary infrastructure
Qdrant
$25+/month
- Cloud hosting required
- Self-hosted option available
- Infrastructure management overhead
Weaviate
$25+/month
- Kubernetes deployment typical
- Cloud provider costs
- Ops complexity
Cost Breakdown: VecLabs Architecture
| Component | Cost | Details | |-----------|------|---------|| | Vector Storage | ~0.000039 per MB per epoch) | | Merkle Root | ~0 | Rust binary runs on your infrastructure | | Data Ownership | Priceless | Encrypted with your Solana wallet—VecLabs cannot read your data |Benchmark Methodology
Test Configuration
Hardware Specifications
Test Machine
- CPU: Apple M2 (8-core)
- RAM: 16GB unified memory
- OS: macOS Sonoma
- Rust: 1.85+ with release optimizations
Benchmark Tool
- Framework: Criterion.rs
- Samples: 100 iterations per benchmark
- Warmup: 10 iterations
- Statistical analysis: Included
Distance Functions
Performance by similarity metric (384 dimensions): | Metric | Operation | Latency | |--------|-----------|---------|| | Cosine Similarity | Dot product + normalization | ~50ns | | Euclidean Distance | L2 norm | ~45ns | | Dot Product | Raw vector multiplication | ~40ns |Running Benchmarks Yourself
Install Dependencies
Run Full Benchmark Suite
Custom Benchmark
Reproducibility
Benchmark source code:
benchmarks/hnsw_bench.rsWhy Rust Matters for Vector Search
The Garbage Collection Problem
Python and Go-based vector databases suffer from unpredictable GC pauses that spike latency during high query loads:Production Impact
Python/Go Vector DBs
- ✗ Unpredictable latency spikes
- ✗ GC pauses under concurrent load
- ✗ Variable p99 latency
- ✗ Degrades with scale
Rust VecLabs
- ✓ Consistent sub-5ms p99
- ✓ Zero GC pauses
- ✓ Predictable latency at scale
- ✓ Production-ready performance
Competitor Analysis
Pinecone
- Strength: Managed service, zero ops
- Weakness: 5.8x more expensive, slower queries, proprietary lock-in
- Use Case: Teams prioritizing convenience over cost/performance
Qdrant
- Strength: Also built in Rust, open source
- Weakness: No on-chain verification, centralized storage, 3x VecLabs cost
- Use Case: Self-hosted vector search without blockchain requirements
Weaviate
- Strength: GraphQL API, strong community
- Weakness: Go-based (GC pauses), slowest query times, complex ops
- Use Case: Teams needing semantic search with graph capabilities
VecLabs
- Strength: Fastest queries + on-chain verification + 88% cost reduction
- Weakness: Alpha software, limited ecosystem integrations (for now)
- Use Case: Production AI agents requiring verifiable, high-performance memory
Next Steps
Get Started
Install VecLabs and run your first query in under 5 minutes
Architecture Deep Dive
Learn how Rust HNSW + Solana + Shadow Drive work together
Migration Guide
Migrate from Pinecone to VecLabs in 30 minutes
Cost Calculator
Estimate your monthly VecLabs costs