R-Mem
Long-term memory for AI agents — in Rust
A lightweight study of mem0's memory architecture. Single binary. SQLite-backed. No Python.
3.6 MB binary · 2,826 lines of Rust · < 10 MB RAM · SQLite only · MCP ready · LongMemEval 48.2%
Quick Start · How It Works · Usage · MCP · Performance · Architecture · Roadmap
[!NOTE] This project reimplements mem0's elegant memory architecture in Rust as a learning exercise. Full credit to the mem0 team for the original design. This is not a replacement — it's a study of their approach using a different language. Discussions, ideas, and contributions are welcome!
Why R-Mem?
mem0 is a well-designed memory system with a rich plugin ecosystem. R-Mem asks a narrower question: what if we rewrite just the core memory logic in Rust, backed entirely by SQLite?
The result is the same three-tier architecture — vector memory, graph memory, history — plus a tiered archive system, in 2,826 lines of Rust. No external services. One binary. The trade-off is clear: far fewer integrations, but near-zero operational overhead.
R-Mem was born out of RustClaw — our minimalist Rust AI agent framework. RustClaw needed a memory layer that matched its philosophy: single binary, zero external services. So we studied mem0's architecture and rebuilt it in Rust.
mem0's numbers reflect its richer ecosystem — more stores, more integrations, more flexibility. R-Mem intentionally trades that for a minimal footprint.
What R-Mem adds beyond mem0
| Feature | R-Mem | mem0 |
|---|---|---|
| Tiered Archive | Deleted/updated memories preserved + fallback search | Gone when deleted |
| FTS5 Pre-filter | Two-stage search: keyword → vector (19x faster) | Vector-only |
| MCP Server | Built-in, rustmem mcp for Claude Code / Cursor |
Not available |
| Zero-dependency deploy | Single binary, SQLite, no Docker | Python + pip + vector DB + graph DB |
| Anthropic native | Direct Claude API support | Via OpenAI-compatible proxy |
| Configurable pipeline | [memory] section: thresholds, limits, all tunable |
Hardcoded defaults |
| Memory categories | Auto-classified: preference, personal, plan, professional, health | Unstructured |
🔍 How It Works
Input text
│
├─ 📦 Vector Memory ──────────────────────────────────
│ │
│ ├─ LLM extracts facts
│ │ → ["Name is Alice", "Works at Google"]
│ │
│ ├─ Embedding → cosine similarity search
│ │ (FTS5 pre-filter + vector ranking)
│ │
│ ├─ Integer ID mapping
│ │ (prevents LLM UUID hallucination)
│ │
│ ├─ LLM decides per fact:
│ │ ├─ ADD new information
│ │ ├─ UPDATE more specific → old version archived
│ │ ├─ DELETE contradiction → old version archived
│ │ └─ NONE duplicate — skip
│ │
│ └─ Execute actions + write history
│
├─ 🕸️ Graph Memory ──────────────────────────────────
│ │
│ ├─ LLM extracts entities + relations
│ ├─ Conflict detection (soft-delete old, add new)
│ └─ Multi-value vs single-value handling
│
└─ 🗄️ Archive ───────────────────────────────────────
│
├─ Deleted/superseded memories preserved with embeddings
├─ Fallback search when active results are weak
└─ Auto-compaction when archive exceeds threshold
🚀 Quick Start
Prerequisites
| Requirement | Install |
|---|---|
| Rust 1.75+ | curl --proto '=https' --tlsv1.2 -sSf https://sh.rustup.rs | sh |
| LLM backend | Ollama, OpenAI, or Anthropic |
Install
Or build from source:
&&
# → target/release/rustmem (3.6 MB)
Configure
Create rustmem.toml in the project root:
[]
= "openai"
= "http://127.0.0.1:11434"
= "qwen2.5:32b"
[]
= "openai"
= "http://127.0.0.1:11434"
= "nomic-embed-text"
[]
= "openai"
= "sk-..."
= "gpt-4o"
[]
= "openai"
= "sk-..."
= "text-embedding-3-small"
[]
= "anthropic"
= "sk-ant-..."
= "claude-sonnet-4-6"
[]
= "openai"
= "sk-..."
= "text-embedding-3-small"
Note: Anthropic does not provide embedding models, so
[embedding]uses OpenAI or Ollama even when[llm]uses Anthropic.
Security: R-Mem binds to
127.0.0.1by default (localhost only). Never put API keys in code — userustmem.toml(gitignored) or environment variables (RUSTMEM__LLM__API_KEY).
📖 Usage
CLI
# Add memories
# Semantic search
# List all memories for a user
# Show graph relations
# Start REST API server
REST API
Start with rustmem server, then:
# ➕ Add memory
# 🔍 Search
# 📋 List all
# 🏷️ Filter by category (preference, personal, plan, professional, health, misc)
&category=preference
# 🗑️ Delete
# 📜 History
# 🗄️ View archived memories
# 🕸️ View graph relations
Drop-in for AI Agents
# mem0 (before)
=
# R-Mem (after — just switch to HTTP)
🔌 MCP Server
R-Mem works as an MCP server — give Claude Code or Cursor long-term memory with one command:
# Claude Code
# Cursor (.cursor/mcp.json)
{
}
7 tools available: add_memory, search_memory, list_memories, get_memory, delete_memory, get_graph, reset_memories
⚡ Performance
Benchmarked on Apple Silicon with 10,000 memories (768-dim embeddings):
| Operation | Time | Notes |
|---|---|---|
| Write | 36 µs/record | 10K records in 360ms |
| Brute-force search | 35.8 ms | Scans all 10K embeddings |
| FTS5 + vector search | 1.9 ms | 19x faster — pre-filters then re-ranks |
| Concurrent reads | 2.4 ms/thread | 10 threads, WAL mode, no blocking |
| Storage | 4.2 KB/memory | 10K memories = 40 MB |
Run the benchmark yourself:
LongMemEval
LongMemEval (ICLR 2025) — 500 questions testing long-term memory across 5 capabilities:
| System | Score | Notes |
|---|---|---|
| agentmemory | 96.2% | RAG (stores raw text) |
| MemLayer | 94.4% | RAG (layered index) |
| Zep | 63.8% | RAG + summary |
| mem0 | ~49% | Fact extraction (gpt-4o) |
| R-Mem | 48.2% | Fact extraction (gpt-4o-mini) |
R-Mem nearly matches mem0 using a 20x cheaper model. The gap vs RAG systems is architectural — R-Mem extracts and deduplicates facts rather than storing raw text, which trades verbatim recall for efficient long-term knowledge management.
🏗️ Architecture
src/
├── main.rs CLI entry point (clap)
├── config.rs TOML + env var config
├── server.rs REST API (axum)
├── mcp.rs MCP server (rmcp) — 7 tools over stdio
├── memory.rs Core orchestrator — tiered memory pipeline
├── extract.rs LLM calls: OpenAI + Anthropic native
├── embedding.rs OpenAI-compatible embedding client
├── store.rs SQLite vector store (WAL + FTS5 + archive)
└── graph.rs SQLite graph store (soft-delete, multi-value)
9 files. 2,826 lines. 3.6 MB binary. Zero external services.
🗺️ Roadmap
| Status | Feature | Description |
|---|---|---|
| ✅ | MCP Server | rustmem mcp — 7 tools over stdio for Claude Code / Cursor |
| ✅ | Tiered Archive | Deleted/updated memories preserved + fallback search |
| ✅ | Anthropic Native | Direct Claude API support (no proxy needed) |
| ✅ | FTS5 Index | Full-text pre-filtering for faster search |
| 🔲 | Batch Import | Load existing mem0 exports |
| 🔲 | Multi-modal | Image / audio memory support |
| 🔲 | Agent SDK | Rust crate for direct embedding (no HTTP) |
| 🔲 | Dashboard | Lightweight web UI for memory inspection |
Community contributions welcome — open an issue or PR.
MIT License · v0.3.0
Created by Ad Huang with Claude Code