What is Memori?

Memori is a memory layer for LLM applications, agents, and copilots. It continuously captures interactions, extracts structured knowledge, and intelligently ranks, decays, and retrieves the relevant memories. So your AI remembers the right things at the right time across every session.

Why Memori Cloud?

With the Memori Cloud platform: app.memorilabs.ai, you skip all database configuration. Sign up, get a Memori API key, and start building AI agents with memory in minutes.

LLM Provider Support

OpenAI, Anthropic, Gemini, and Grok (xAI) via direct SDK wrappers. Bedrock is supported through LangChain ChatBedrock. Supports sync, async, streamed, and unstreamed modes, plus LangChain, Agno, and Pydantic AI.

Zero Configuration

No database setup needed. Connect your LLM client with an API key and start building memories immediately.

Framework Integration

Native support for LangChain, Agno, and Pydantic AI with seamless integration into your existing workflows.

Advanced Augmentation

Background AI processing extracts facts, preferences, and relationships from your conversations automatically.

Semantic Recall

Semantic search surfaces the right memories at the right time. Memories are ranked by relevance and importance, with intelligent decay so older or less relevant facts recede — so your AI stays contextually aware without clutter. Use manual recall when you need to display memories in your UI, build custom prompts, or debug. See How Memori Works for automatic vs manual recall and tuning.

from memori import Memori
from openai import OpenAI

# Requires MEMORI_API_KEY and OPENAI_API_KEY in your environment
client = OpenAI()
mem = Memori().llm.register(client)

# Track conversations by user and process
mem.attribution(entity_id="user_123", process_id="support_agent")

# All conversations automatically persisted and recalled
response = client.chat.completions.create(
    model="gpt-4o-mini",
    messages=[{"role": "user", "content": "My favorite color is blue."}]
)

Core Concepts

ConceptDescriptionExample
EntityPerson, place, or thing (like a user)entity_id="user_123"
ProcessYour agent, LLM interaction, or programprocess_id="support_agent"
SessionGroups LLM interactions togetherAuto-generated UUID, manually manageable
AugmentationBackground AI enhancement of memoriesAuto-runs after wrapped LLM calls
RecallRetrieve relevant memories from previous interactionsAuto-injects recalled memories

Architecture Overview

The diagram has three lanes: your app, the Memori SDK, and Memori Cloud. Your app calls the LLM normally, Memori captures context on the response path, and memory processing continues in the background.

Synchronous capture: conversation messages are sent to Memori Cloud while your normal LLM flow continues.

Recall injection: relevant memories are fetched from managed storage and injected into later prompts.

Async augmentation: Memori Cloud extracts facts, preferences, rules, events, and relationships from conversations without blocking your app.

"Memori Cloud architecture"