The First AI Memory That Thinks Like a Brain
5-layer cognitive architecture. Zero-LLM ingestion pipeline. Self-improving memory that decays, consolidates, and reasons — like a real brain. The memory OS that every other system forgot to build.
Not a Demo. Not a Roadmap. Running Right Now.
Mnemosyne powers a 10-machine AI fleet with sub-200ms retrieval latency. Every feature on this page is in production.
AI Agents Have Amnesia. Mnemosyne Gives Them a Brain.
Every Other Memory System
- Burns LLM tokens on every memory stored (~$0.01 each)
- Retrieves by single signal only (cosine similarity)
- No knowledge graph, or paywalled at $249/mo
- Single-agent only — no multi-agent collaboration
- Static storage — quality degrades over time
- No activation decay, no consolidation, no reasoning
Mnemosyne
- Memories persist across sessions, restarts, and context window resets
- Retrieval uses 5 signals weighted by detected intent — not just cosine similarity
- Knowledge grows through auto-linking, graph expansion, and cross-agent corroboration
- Quality improves via reinforcement learning and autonomous 4-phase consolidation
- Agents collaborate through a real-time memory mesh with pub/sub broadcasting
- Zero LLM calls during ingestion — deterministic, $0 per memory, works offline
5 Lines to Cognitive Memory
Drop-in TypeScript SDK. Zero LLM required for the ingestion pipeline. Your agent has a brain in under a minute.
Works With Your Stack
Mnemosyne is the brain. The databases are just where data lives.
5-Layer Cognitive Architecture
Inspired by how the human brain organizes, retrieves, and strengthens memories over time. Each layer is independently toggleable.
Self-Improvement
Reinforcement Learning, Active Consolidation, Flash Reasoning, Theory of Mind
Cognitive
Activation Decay, Multi-Signal Scoring, Intent-Aware Retrieval, Diversity Reranking
Knowledge Graph
Temporal Graph, Auto-Linking, Path Finding, Timeline Reconstruction
Pipeline
12-Step Zero-LLM Ingestion: Security, Embed, Dedup, Extract, Classify, Score, Link, Graph, Broadcast
Infrastructure
Pluggable Storage Backends: Vector DB, Graph DB, 2-Tier Cache, Pub/Sub — bring your own or use defaults
Every Feature. All Shipping.
33 production features across 5 cognitive layers. Every feature is independently toggleable — start simple, enable progressively.
Infrastructure
6 featuresVector Storage
768-dim embeddings with HNSW indexing, sub-linear scaling to billions. Supports Qdrant, Pinecone, Weaviate, Chroma
2-Tier Cache
L1 in-memory (50 entries, 5min TTL) + L2 cache (1hr TTL) for sub-10ms recall. Supports Redis, Memcached, or in-memory
Pub/Sub Broadcast
Real-time memory events across your entire agent mesh via configurable pub/sub backend
Knowledge Graph
Temporal entity graph with auto-linking, path finding, timeline reconstruction. Supports FalkorDB, Neo4j, or disable
Bi-Temporal Model
Every memory tracks eventTime (when it happened) + ingestedAt (when stored)
Soft-Delete Architecture
Memories are never physically deleted — full audit trails and recovery
Pipeline
6 features12-Step Ingestion
Security → embed → dedup → extract → classify → score → link → graph → broadcast
Zero-LLM Pipeline
All classification, extraction, scoring runs algorithmically — $0 per memory, <50ms
Security Filter
3-tier classification (public/private/secret), blocks API keys and credentials
Smart Dedup & Merge
Cosine ≥0.92 = duplicate merge, 0.70–0.92 = conflict alert broadcast
Entity Extraction
Automatic identification of people, machines, IPs, dates, technologies, URLs — zero LLM
7-Type Taxonomy
Episodic, semantic, preference, relationship, procedural, profile, core — algorithmic
Knowledge Graph
5 featuresTemporal Queries
"What was X connected to as of date Y?" — relationships carry since timestamps
Auto-Linking
New memories automatically discover and link to related memories, bidirectional, Zettelkasten-style
Path Finding
Shortest-path queries between any two entities with configurable max depth
Timeline Reconstruction
Ordered history of all memories mentioning a given entity
Depth-Limited Traversal
Configurable graph exploration (default: 2 hops) balancing relevance vs. noise
Cognitive
6 featuresActivation Decay
Logarithmic decay model — critical memories stay months, core/procedural are immune
Multi-Signal Scoring
5 signals: similarity, recency, importance×confidence, frequency, type relevance
Intent-Aware Retrieval
Auto-detects query intent (factual, temporal, procedural, preference, exploratory)
Diversity Reranking
Cluster detection, overlap penalty, type diversity — prevents echo chambers in results
4-Tier Confidence
Mesh Fact ≥0.85, Grounded 0.65–0.84, Inferred 0.40–0.64, Uncertain <0.40
Priority Scoring
Urgency × Domain composite score — critical+technical = 1.0, background+general = 0.2
Self-Improvement
10 featuresReinforcement Learning
Feedback loop tracks usefulness, auto-promotes memories with >0.7 ratio after 3+ retrievals
Active Consolidation
4-phase autonomous maintenance: contradiction detection, dedup merge, promotion, demotion
Flash Reasoning
BFS traversal through linked memory graphs, reconstructs multi-step logic chains
Theory of Mind (TOMA)
"What does Agent-B know about X?" — knowledge gap analysis across the mesh
Cross-Agent Synthesis
3+ agents agree on a fact → auto-synthesized into fleet-level insight
Proactive Recall
Generates speculative queries from incoming prompts, injects context before the agent asks
Session Survival
Snapshot/recovery across context window resets — zero discontinuity
Observational Memory
Compresses raw conversation streams into structured, high-signal memory cells
Procedural Memory
Learned procedures as first-class objects, immune to decay, shared across the mesh
Mesh Sync
Named, versioned shared state blocks with real-time broadcast propagation
10 Capabilities From Research Papers. All Production-Ready.
These capabilities exist almost exclusively in academic literature and closed research labs. Mnemosyne ships all 10 as deployable infrastructure.
| Capability | Industry Status | Mnemosyne |
|---|---|---|
Flash Reasoning Chain-of-thought traversal through linked memory graphs with BFS and cycle detection | Research paper only | Production |
Theory of Mind for Agents Agents model what other agents know — enabling task routing and collaborative problem-solving | Research paper only | Production |
Observational Memory Raw conversation streams compressed into structured observations, like human working memory | Research paper only | Production |
Reinforcement Learning on Memory Feedback loop auto-promotes useful memories and flags misleading ones | Research paper only | Production |
Self-Improving Consolidation 4-phase autonomous maintenance: contradictions, dedup, promotion, demotion | Not implemented anywhere | Production |
Cross-Agent Cognitive State Named, versioned shared blocks participate in retrieval, reasoning, and consolidation | Not implemented anywhere | Production |
Bi-Temporal Knowledge Graph Tracks what was true at any point in time — eventTime + ingestedAt on every relationship | Research paper only | Production |
Proactive Anticipatory Recall Speculative queries surface relevant context before the agent asks for it | Not implemented anywhere | Production |
Procedural Memory / Skill Library Learned procedures as first-class objects, immune to decay, shared across the mesh | Not implemented anywhere | Production |
Session Survival Cognitive continuity across context window resets via snapshot/recovery | Not implemented anywhere | Production |
33 Features. 28 That Nobody Else Has.
Every feature listed is in production. Not planned. Not in beta. Shipping.
| Feature | Mnemosyne | Mem0 | Zep | Cognee | LangMem | Letta |
|---|---|---|---|---|---|---|
| Pipeline & Ingestion | ||||||
| Zero-LLM Ingestion Pipeline | ||||||
| 12-Step Structured Pipeline | ||||||
| Security Filter (Secret Blocking) | ||||||
| Smart Dedup with Semantic Merge | ||||||
| Conflict Detection & Alerts | ||||||
| 7-Type Memory Taxonomy | ||||||
| Entity Extraction (Zero-LLM) | LLM $ | LLM $ | LLM $ | |||
| Cognitive Features | ||||||
| Activation Decay Model | ||||||
| Multi-Signal Scoring (5 Signals) | ||||||
| Intent-Aware Retrieval | ||||||
| Diversity Reranking | ||||||
| Flash Reasoning Chains | ||||||
| Reinforcement Learning | ||||||
| Active Consolidation (4-Phase) | ||||||
| Proactive Recall | ||||||
| Session Survival (Compaction) | ||||||
| Observational Memory | ||||||
| Preference Modeling | ||||||
| Knowledge Graph | ||||||
| Built-in Knowledge Graph | $249/mo | |||||
| Temporal Graph Queries | ||||||
| Auto-Linking (Bidirectional) | ||||||
| Path Finding Between Entities | ||||||
| Timeline Reconstruction | ||||||
| Bi-Temporal Data Model | ||||||
| Multi-Agent | ||||||
| Real-Time Broadcast (Pub/Sub) | ||||||
| Theory of Mind (TOMA) | ||||||
| Cross-Agent Synthesis | ||||||
| Knowledge Gap Analysis | ||||||
| Shared State Blocks (Mesh Sync) | ||||||
| Infrastructure | ||||||
| 2-Tier Caching (L1 + L2) | ||||||
| Soft-Delete Architecture | ||||||
| Procedural Memory (Skill Library) | ||||||
| CLI Tools | ||||||
MIT License. Free Knowledge Graph. $0 Per Memory.
Competitors paywall their graph, charge per memory via LLM, and gate multi-agent behind enterprise tiers.
| Mnemosyne | Mem0 | Zep | Letta | |
|---|---|---|---|---|
| License | MIT (fully open) | Open core (limited) | Open core (limited) | Open source |
| Self-hosted | Free — all features | Free — limited features | Free — limited features | Free |
| Knowledge graph | Free (self-hosted) | $249/mo (Pro tier) | N/A | N/A |
| Per memory stored | $0.00 (zero LLM) | ~$0.01 (LLM call) | ~$0.01 (LLM call) | ~$0.01 (LLM call) |
| 100K memories cost | $0 | ~$1,000 | ~$1,000 | ~$1,000 |
| Multi-agent | Free — built-in mesh | Enterprise pricing | N/A | N/A |
| Cognitive features | Free — all 10 | N/A | N/A | Session mgmt only |
LLM costs estimated at ~$0.01/memory (conservative). Mnemosyne's zero-LLM pipeline has exactly $0 in per-memory costs beyond infrastructure.
Store. Recall. Learn. Collaborate.
Four operations that transform stateless agents into intelligent, collaborative systems.
Store
Input goes through a 12-step zero-LLM pipeline: security filter, embedding, dedup, extraction, classification, scoring, linking, graph ingest, and broadcast. All in <50ms.
Recall
Queries hit the 2-tier cache, then vector search with 5-signal intent-aware scoring, diversity reranking, graph enrichment, and flash reasoning chains.
Learn
Feedback signals promote useful memories and flag poor ones. Active consolidation merges duplicates, resolves contradictions, and promotes popular knowledge.
Collaborate
Agents share memories via pub/sub mesh. Theory of Mind queries what others know. 3+ agents agreeing on a fact triggers fleet-level synthesis.
10-40x Faster Ingestion Than Any Competitor
Real numbers from a 10-machine AI fleet running Mnemosyne in production.
| Operation | Mnemosyne | LLM-Based Systems |
|---|---|---|
| Store (full pipeline) | <50ms | 500ms – 2s |
| Recall (cached) | <10ms | No caching |
| Recall (uncached) | <200ms | 200ms – 500ms |
| Consolidation | ~1,000/min | Not available |
| Embedding generation | ~15ms (cached) | ~15ms |
Built for Agents That Actually Think
From single-agent coding assistants to enterprise-scale agent meshes.
AI Coding Assistants
Session survival, procedural memory, temporal graph. Agents remember project context, deployment procedures, and past debugging sessions.
Enterprise Knowledge Agents
Agent mesh, Theory of Mind, cross-agent synthesis. Specialized agents (HR, IT, Finance) build domain expertise while sharing verified facts.
Customer Support
Preference tracking, reinforcement learning, procedural memory. Agents remember customer history, resolution patterns, and preferences.
Research Assistants
Flash reasoning, auto-linking, knowledge graph. Agents accumulate domain knowledge and surface non-obvious connections between findings.
DevOps & Infrastructure
Temporal queries, proactive warnings, mesh sync. Agents remember topology, incidents, and answer "What changed since last stable state?"
Personal AI Companions
Activation decay, observational memory, preference modeling. Long-running assistants that develop genuine understanding over time.
9 Tools. Any Agent Framework.
Drop-in tools that work with Claude, GPT, LangChain, CrewAI, AutoGen, or any LLM agent framework.
memory_recallMulti-signal search with intent detection, diversity reranking, graph enrichment, flash reasoning
memory_storeFull 12-step ingestion: security filter, dedup, classify, link, graph ingest, broadcast
memory_forgetSoft-delete by ID or semantic search, mesh-wide cache invalidation
memory_block_getRead a named shared memory block (Mesh Sync)
memory_block_setWrite/update a named shared block with versioning and broadcast
memory_feedbackReinforcement signal — drives memory promotion/demotion
memory_consolidate4-phase active maintenance: contradictions, dedup, promotion, demotion
memory_tomaQuery what a specific agent knows about a topic (Theory of Mind)
before_agent_startAutomatic hook: session recovery, proactive recall, context injection
Give Your Agents the Memory They Deserve
Open source. MIT licensed. 33 features. $0 per memory. Deploy cognitive memory in minutes.
Because intelligence without memory isn't intelligence.