Skip to main content

What is Memori Cloud?

Memori is a memory layer for LLM applications, agents, and copilots. It continuously captures interactions, extracts structured knowledge, and intelligently ranks, decays, and retrieves the relevant memories. So your AI remembers the right things at the right time across every session.

Why Choose Memori Cloud?

With the Memori Cloud platform at app.memorilabs.ai, you skip all database configuration. Sign up, get a Memori API key, and start building AI agents with memory in minutes.

Zero Configuration

No database setup needed. Connect your LLM client with an API key and start building memories immediately.
from memori import Memori
from openai import OpenAI

# Requires MEMORI_API_KEY and OPENAI_API_KEY in your environment
client = OpenAI()
mem = Memori().llm.register(client)

# Track conversations by user and process
mem.attribution(entity_id="user_123", process_id="support_agent")

# All conversations automatically persisted and recalled
response = client.chat.completions.create(
    model="gpt-4o-mini",
    messages=[{"role": "user", "content": "My favorite color is blue."}]
)

LLM Provider Support

Memori Cloud supports major LLM providers with sync, async, streamed, and unstreamed modes:
  • OpenAI - Chat Completions & Responses API
  • Anthropic - Claude models
  • Google Gemini - Gemini models
  • Grok (xAI) - Grok models
  • AWS Bedrock - via LangChain ChatBedrock

Framework Integration

Native support for popular AI frameworks:
  • LangChain - Seamless integration with LangChain chains
  • Agno - Built-in Agno support
  • Pydantic AI - Type-safe AI development

Advanced Augmentation

Background AI processing extracts facts, preferences, and relationships from your conversations automatically:
  • Attributes - User characteristics and properties
  • Events - Important occurrences and milestones
  • Facts - Verified information and details
  • Preferences - User likes and dislikes
  • Relationships - Connections between entities
  • Rules - Behavioral patterns and constraints
  • Skills - Capabilities and expertise

Semantic Recall

Semantic search surfaces the right memories at the right time. Memories are ranked by relevance and importance, with intelligent decay so older or less relevant facts recede — so your AI stays contextually aware without clutter.
Use manual recall when you need to display memories in your UI, build custom prompts, or debug. Automatic recall happens seamlessly during LLM calls.

Core Concepts

ConceptDescriptionExample
EntityPerson, place, or thing (like a user)entity_id="user_123"
ProcessYour agent, LLM interaction, or programprocess_id="support_agent"
SessionGroups LLM interactions togetherAuto-generated UUID, manually manageable
AugmentationBackground AI enhancement of memoriesAuto-runs after wrapped LLM calls
RecallRetrieve relevant memories from previous interactionsAuto-injects recalled memories

Architecture Overview

The diagram shows three lanes: your app, the Memori SDK, and Memori Cloud. Your app calls the LLM normally, Memori captures context on the response path, and memory processing continues in the background. Synchronous capture: Conversation messages are sent to Memori Cloud while your normal LLM flow continues. Recall injection: Relevant memories are fetched from managed storage and injected into later prompts. Async augmentation: Memori Cloud extracts facts, preferences, rules, events, and relationships from conversations without blocking your app. Memori Cloud architecture

Memory Usage Plans

PlanMemories Created / monthMemories Recalled / month
Free (with API key)5,00015,000
Starter25,000100,000
Pro150,000500,000
EnterpriseCustomCustom

Next Steps

1

Install Memori

Get started by installing the Memori SDK and setting up your API key.Go to Installation →
2

Quick Start

Build your first memory-enabled application in under 3 minutes.Go to Quick Start →
3

Explore the Dashboard

Manage API keys, test memories, and monitor usage in the dashboard.Go to Dashboard →

Build docs developers (and LLMs) love