The State of Agent Memory in 2026
We audited 10 open-source agent memory projects — 120K+ GitHub stars, $31.5M in funding — to map where the field actually stands. Here's what we found.
The number that started everything
Mem0's npm package — mem0ai — was getting 3,000+ weekly downloads before it even had a working JavaScript SDK. The README pointed to a Python library. People were installing an empty package just because the word "memory" was in the name.
That one metric convinced us to look deeper. If there's this much demand for agent memory, what does the supply actually look like?
By the numbers
We audited 10 open-source agent memory projects spanning three distinct paradigms: vector-based retrieval, knowledge graphs, and hybrid approaches. Collectively, they represent 120,000+ GitHub stars, $31.5M in venture funding, and wildly different bets on how agents should remember.
Key findings
-
Graphiti's bi-temporal model is the most architecturally sophisticated approach we found. It tracks both when an event happened and when the system learned about it — essential for correcting stale memories without losing history.
-
memU's salience scoring borrows from cognitive science. Memories decay based on recency, frequency, and emotional weight. It's the closest anyone has come to modeling how human memory actually prioritizes information.
-
SimpleMem's multi-query retrieval generates multiple search queries from a single user prompt, then merges results. Simple idea, significant accuracy improvement over single-vector search.
-
Mem0 fixed a UUID hallucination bug where the LLM was generating fake memory IDs during graph updates. Their solution: strip UUIDs from the prompt entirely and let the system handle references. A reminder that LLMs and structured identifiers don't mix.
-
mcp-memory-service implements exponential decay with a configurable half-life. Memories literally fade unless reinforced — the most aggressive temporal model in the landscape.
-
Letta (formerly MemGPT) takes the most literal approach: a full operating-system metaphor with main memory, archival storage, and an agent that manages its own context window.
The comparison
| Project | Approach | Retrieval | Memory Evolution | Real-time Signals |
|---|---|---|---|---|
| Mem0 | Hybrid (vector + graph) | Semantic search | LLM-driven updates | None |
| Graphiti | Knowledge graph | Bi-temporal queries | Episodic + semantic layers | None |
| Letta | OS metaphor | Agent-managed context | Self-editing memory blocks | None |
| memU | Cognitive model | Salience-weighted | Decay + reinforcement | None |
| SimpleMem | Vector store | Multi-query expansion | Append-only | None |
| mcp-memory-service | Vector + decay | Similarity search | Exponential decay | None |
| A-MEM | Autonomous notes | Reflection-triggered | Self-organizing | None |
| Basic Memory | File-based | Markdown search | Manual editing | None |
| Cognee | Graph + vector | Hybrid retrieval | Pipeline-driven | None |
| Zep | Temporal graph | Fact extraction | Continuous updates | None |
The missing layer
Notice the last column. Every project in this landscape is optimized for conversational context — chat history, user preferences, accumulated facts from dialogue. None of them are ingesting real-time engineering signals: commits landing, PRs being reviewed, deployments shipping, incidents resolving.
That's the gap. Your codebase generates a continuous stream of structured events that carry institutional knowledge — why a decision was made, what broke last time, which patterns the team has converged on. No memory framework is capturing it.
This is what we're building with Mneme: a memory layer that starts from real-time signals, not chat transcripts. Architecture decisions extracted from PR reviews. Failure patterns learned from incident timelines. Tribal knowledge surfaced from the code itself.
Get the full report
Enter your email to download the complete 15 min read deep dive as a PDF.
Related research
Why Every AI Coding Tool Is a Perpetual New Hire
84% of developers use AI coding tools. Only 33% trust the output. Adoption is accelerating and trust is collapsing simultaneously — here's why.
We Deleted a Feature 4 Days After Shipping It
We built a memory system for AI coding tools, shipped a feature that remembered too much, and deleted it when retrieval quality degraded. Here's what forgetting taught us.