Installation
This guide covers everything you need to install and configure REMem for your use case.Requirements
Python version
REMem requires Python 3.10 or higher. You can check your Python version:REMem has been tested with Python 3.10, 3.11, and 3.12. We recommend using Python 3.10+ for the best compatibility.
System requirements
- Memory: At least 8GB RAM (16GB+ recommended for large datasets)
- GPU: Optional but recommended for faster embedding generation and offline LLM inference
- Storage: Varies based on dataset size and caching
Installation methods
Install from source (recommended)
Clone the repository and install in editable mode:This installs REMem along with all required dependencies from
pyproject.toml.Core dependencies
REMem automatically installs these core dependencies:Graph and numerical computation
- networkx (3.4.2) — Graph algorithms and structures
- python_igraph (0.11.8) — Fast graph operations
- numpy (1.26.4) — Numerical computing
- scipy (1.14.1) — Scientific computing
Machine learning and embeddings
- torch (2.6.0) — PyTorch for deep learning
- sentence_transformers (3.3.1) — Embedding model interface
- transformers (4.51.1) — Hugging Face transformers
- nano_vectordb (0.0.4.3) — Lightweight vector database
LLM integration
- openai (≥1.0.0) — OpenAI API client
- vllm (0.8.5post1) — Offline LLM inference
- dspy (2.5.29) — DSPy for prompt optimization
- tiktoken (0.7.0) — Token counting
Utilities
- tqdm (4.66.6) — Progress bars
- tenacity (8.5.0) — Retry logic
- pydantic (2.10.4) — Data validation
- pandas — Data manipulation
- nltk — Natural language processing
See
pyproject.toml in the repository for the complete list of dependencies and version constraints.API keys and configuration
OpenAI API
For online mode with OpenAI models:Azure OpenAI
For Azure OpenAI deployments:Environment variables
REMem respects these environment variables:Embedding models
REMem supports multiple embedding models:NV-Embed-v2 (recommended)
OpenAI embeddings
GritLM
Qwen3
Local embedding models (NV-Embed-v2, GritLM, Qwen3) will be downloaded from Hugging Face on first use. Make sure you have sufficient disk space.
LLM backends
Online mode (OpenAI API)
Default mode for development and smaller workloads:gpt-4o-mini(recommended for cost-effectiveness)gpt-4ogpt-3.5-turbo
Offline mode (vLLM)
For batch processing and local inference:Verify installation
Run this simple script to verify your installation:Troubleshooting
Import errors
If you encounter import errors, make sure all dependencies are installed:CUDA/GPU issues
If you have GPU issues, ensure PyTorch is installed correctly:Embedding model download
If embedding models fail to download, you can pre-download them:Out of memory errors
For large datasets, reduce batch sizes:Development setup
For development, install additional tools:- Black for code formatting (line length: 120)
- Ruff for linting (compatible with Python 3.10+)
Next steps
Quickstart
Build your first REMem application
Configuration
Learn about advanced configuration options
Benchmarks
Run REMem on research benchmarks
Examples
Browse complete code examples