mentisdb 0.7.1

Semantic, hash-chained durable memory for long-running and multi-agent systems.
Documentation

MentisDB

MentisDB is a durable semantic memory engine and versioned skill registry for AI agents — a persistent, hash-chained brain that survives context resets, model swaps, and team turnover.

It stores semantically typed thoughts in an append-only, hash-chained memory log through a swappable storage adapter layer. The skill registry is a git-like immutable version store for agent instruction bundles — every upload is a new version, history is never overwritten, and every version is cryptographically signable.


Why MentisDB

Harness Swapping — the same durable memory works across every AI coding environment. Connect Claude Code, OpenAI Codex, GitHub Copilot CLI, Qwen Code, Cursor, VS Code, or any MCP-capable host to the same mentisdbd daemon and your agents share one brain, regardless of which tool you picked up today.

Zero Knowledge Loss Across Context Boundaries — when an agent's context window fills, it writes a Summary checkpoint to MentisDB, compacts, reloads mentisdb_recent_context, and continues without losing a single decision. Chat history is ephemeral. MentisDB is permanent.

Fleet Orchestration at Scale — one project manager agent decomposes work, dispatches a parallel fleet of specialists, each pre-warmed with shared memory, and synthesizes results wave by wave. MentisDB is the coordination substrate: every agent reads from the same chain and writes its lessons back. The fleet's collective intelligence compounds.

Versioned Skill Registry — skills are not just stored, they are versioned like a git repository. Every upload to an existing skill_id creates a new immutable version (stored as a unified diff). Any historical version is reconstructable. Skills can be deprecated or revoked while full audit history is preserved. Uploading agents with registered Ed25519 keys must cryptographically sign their uploads — provenance is verifiable, not assumed.

Session Resurrection — any agent can call mentisdb_recent_context and immediately know exactly where the project stands, what decisions were made, what traps were already hit, and what comes next — without re-reading code, re-running exploratory searches, or asking the human to re-explain context that was earned through hours of work.

Self-Improving Agent Fleets — agents upload updated skill files after learning something new. A skill checked in at the start of a project is better by the end of it. Combine with Ed25519 signing to create a verifiable, tamper-evident record of which agent authored which version of institutional knowledge.

Multi-Agent Shared Brain — multiple agents, multiple roles, multiple owners can write to the same chain key simultaneously. Every thought carries a stable agent_id. Queries filter by agent identity, thought type, role, tags, concepts, importance, and time windows. The chain represents the full collective intelligence of an entire orchestration system, not just one session.

Lessons That Outlive Models — architectural decisions, hard constraints, non-obvious failure modes, and retrospectives written to MentisDB survive chat loss, model upgrades, and team changes. The knowledge compounds instead of evaporating. A new engineer or a new agent boots up, loads the chain, and inherits everything the team learned.


Quick Start

Install the daemon:

cargo install mentisdb

Connect your local AI tools the fast way:

mentisdbd wizard

Or target one integration explicitly:

mentisdbd setup codex
mentisdbd setup all --dry-run

Then start the daemon:

mentisdbd

On an interactive first run with no configured client integrations, mentisdbd offers to launch the setup wizard immediately after startup so you do not have to guess the next command.

Run persistently after closing your SSH session:

nohup mentisdbd &

Modern MCP clients bootstrap themselves from the MCP handshake:

  • initialize.instructions tells the agent to read mentisdb://skill/core
  • resources/read(mentisdb://skill/core) delivers the embedded operating skill
  • GET /mentisdb_skill_md remains available only as a compatibility fallback

If you need to wire a tool manually, here are the raw MCP commands/configs:

# Claude Code
claude mcp add --transport http mentisdb http://127.0.0.1:9471

# OpenAI Codex
codex mcp add mentisdb --url http://127.0.0.1:9471

# Qwen Code
qwen mcp add --transport http mentisdb http://127.0.0.1:9471

# GitHub Copilot CLI — use /mcp add in interactive mode,
# or write ~/.copilot/mcp-config.json manually (see below)

What Is In This Folder

mentisdb/ contains:

  • the standalone mentisdb library crate
  • server support for HTTP MCP and REST, enabled by default
  • the mentisdbd daemon binary
  • dedicated tests under mentisdb/tests

Makefile

A Makefile is included at the repository root. All common workflows have a target:

make build          # fmt + release build
make build-mentisdbd # build only the daemon binary
make release        # fmt, check, clippy, build, test, doc in sequence
make fmt            # cargo fmt
make check          # cargo check (lib + binary)
make clippy         # cargo fmt + clippy --all-targets -D warnings
make test           # cargo test
make bench          # Criterion benchmarks, output tee'd to /tmp/mentisdb_bench_results.txt
make doc            # cargo doc --all-features
make install        # cargo install --path . --locked
make publish        # cargo publish
make publish-dry-run
make clean
make help           # list all targets with descriptions

Build

make build

Or directly with Cargo:

cargo build --release

Build only the library without the default daemon/server stack:

cargo build --no-default-features

Test

make test

Or directly:

cargo test

Run tests for the library-only build:

cargo test --no-default-features

Run rustdoc tests:

cargo test --doc

Benchmarks

MentisDB ships a Criterion benchmark suite and a harness-free HTTP concurrency benchmark:

make bench

Or directly:

cargo bench

Results are also written to /tmp/mentisdb_bench_results.txt so numbers persist across terminal sessions.

Benchmark coverage:

  • benches/thought_chain.rs — 10 benchmarks: append throughput, query latency, traversal patterns
  • benches/search_baseline.rs — 4 benchmarks: lexical/filter-first search baseline over content, registry text, indexed+text intersections, and newest-tail limits
  • benches/search_ranked.rs — 4 benchmarks: additive ranked retrieval over lexical content, filtered ranked queries, and heuristic fallback, plus a baseline append-order comparison
  • benches/skill_registry.rs — 12 benchmarks: skill upload, search, delta reconstruction, lifecycle
  • benches/http_concurrency.rs — starts mentisdbd in-process on a random port; measures write and read throughput at 100 / 1k / 10k concurrent Tokio tasks with p50/p95/p99 latency reporting

Baseline numbers from the DashMap concurrent chain lookup refactor: 750–930 read req/s at 10k concurrent tasks, compared to a sequential bottleneck on the previous RwLock<HashMap> implementation.


Generate Docs

make doc

Or directly:

cargo doc --no-deps

Generate docs for the library-only build:

cargo doc --no-deps --no-default-features

Run The Daemon

The standalone executable is mentisdbd.

Run it from source:

cargo run --bin mentisdbd

Install it from the crate directory:

make install
# or
cargo install --path . --locked

mentisdbd now owns both daemon startup and local integration setup:

mentisdbd setup codex
mentisdbd setup all --dry-run
mentisdbd wizard
mentisdbd

When it starts, it serves:

  • an MCP server
  • a REST server
  • an HTTPS web dashboard

Before serving traffic, it:

  • migrates or reconciles discovered chains to the current schema and default storage adapter
  • verifies chain integrity and attempts repair from valid local sources when possible
  • migrates the skill registry from V1 to V2 format if needed (idempotent; safe to run repeatedly)

Once startup completes, it prints:

  • the active chain directory, default chain key, and bound MCP/REST/dashboard addresses
  • a catalog of all exposed HTTP endpoints with one-line descriptions
  • a per-chain summary with version, adapter, thought count, and per-agent counts

Daemon Configuration

mentisdbd is configured with environment variables:

  • MENTISDB_DIR Directory where MentisDB storage adapters store chain files.
  • MENTISDB_DEFAULT_CHAIN_KEY Default chain_key used when requests omit one. Default: borganism-brain. MENTISDB_DEFAULT_KEY is accepted as a deprecated alias.
  • MENTISDB_STORAGE_ADAPTER Default storage backend for newly created chains. Supported values: binary, jsonl. Default: binary
  • MENTISDB_VERBOSE When unset, verbose interaction logging defaults to true. Supported explicit values: 1, 0, true, false.
  • MENTISDB_LOG_FILE Optional path for interaction logs. When set, MentisDB writes interaction logs to that file even if console verbosity is disabled. If MENTISDB_VERBOSE=true, the same lines are also mirrored to the console logger.
  • MENTISDB_BIND_HOST Bind host for both HTTP servers. Default: 127.0.0.1
  • MENTISDB_MCP_PORT MCP server port. Default: 9471
  • MENTISDB_REST_PORT REST server port. Default: 9472
  • MENTISDB_DASHBOARD_PORT HTTPS dashboard port. Default: 9475. Set to 0 to disable the web dashboard.
  • MENTISDB_DASHBOARD_PIN Optional PIN required to access the dashboard. Leave unset only for trusted localhost use.
  • MENTISDB_AUTO_FLUSH Controls per-write durability of the binary storage adapter.
    • true (default): every append_thought flushes to disk immediately. Full durability.
    • false: writes are batched and flushed every 16 appends (FLUSH_THRESHOLD). Up to 15 thoughts may be lost on a hard crash or power failure, but write throughput increases significantly for multi-agent hubs with many concurrent writers. Supported values: 1, 0, true, false. Has no effect on the jsonl adapter.
  • MENTISDB_UPDATE_CHECK Background GitHub release check for mentisdbd. Enabled by default; set 0, false, no, or off to disable update checks after startup. Default: true
  • MENTISDB_UPDATE_REPO Optional GitHub owner/repo override used by the updater. Default: CloudLLM-ai/mentisdb

Example — full durability (production default):

MENTISDB_DIR=/tmp/mentisdb \
MENTISDB_DEFAULT_CHAIN_KEY=borganism-brain \
MENTISDB_STORAGE_ADAPTER=binary \
MENTISDB_VERBOSE=true \
MENTISDB_LOG_FILE=/tmp/mentisdb/mentisdbd.log \
MENTISDB_BIND_HOST=127.0.0.1 \
MENTISDB_MCP_PORT=9471 \
MENTISDB_REST_PORT=9472 \
MENTISDB_DASHBOARD_PIN=change-me \
MENTISDB_AUTO_FLUSH=true \
cargo run --bin mentisdbd

Example — high-throughput write mode (multi-agent hub):

MENTISDB_DIR=/var/lib/mentisdb \
MENTISDB_AUTO_FLUSH=false \
MENTISDB_BIND_HOST=0.0.0.0 \
mentisdbd

Automatic Update Check

mentisdbd checks GitHub releases in the background after startup and can offer to update itself with cargo install.

  • checks are enabled by default
  • version comparison uses only the first three numeric components, so a tag like 0.6.1.14 is treated as core version 0.6.1
  • interactive terminals get an ASCII prompt window with Y / N
  • non-interactive terminals never block; they print the exact manual cargo install command instead

Disable the automatic check:

MENTISDB_UPDATE_CHECK=0 \
mentisdbd

Server Surfaces

MCP endpoints:

  • GET /health
  • POST /
  • POST /tools/list
  • POST /tools/execute

REST endpoints:

  • GET /health
  • GET /mentisdb_skill_md
  • GET /v1/skills
  • GET /v1/skills/manifest
  • GET /v1/chains
  • POST /v1/bootstrap
  • POST /v1/agents
  • POST /v1/agent
  • POST /v1/agent-registry
  • POST /v1/agents/upsert
  • POST /v1/agents/description
  • POST /v1/agents/aliases
  • POST /v1/agents/keys
  • POST /v1/agents/keys/revoke
  • POST /v1/agents/disable
  • POST /v1/thought
  • POST /v1/thoughts
  • POST /v1/thoughts/genesis
  • POST /v1/thoughts/traverse
  • POST /v1/retrospectives
  • POST /v1/search
  • POST /v1/lexical-search
  • POST /v1/ranked-search
  • POST /v1/context-bundles
  • POST /v1/recent-context
  • POST /v1/memory-markdown
  • POST /v1/skills/upload
  • POST /v1/skills/search
  • POST /v1/skills/read
  • POST /v1/skills/versions
  • POST /v1/skills/deprecate
  • POST /v1/skills/revoke
  • POST /v1/head

Search Semantics

MentisDB keeps its baseline thought search surface filter-first and append-order. Ranked, graph-aware, and vector retrieval are additive surfaces layered on top of that stable baseline.

Today, the main search APIs are:

  • MentisDb::query(&ThoughtQuery)
  • POST /v1/search
  • mentisdb_search

Current behavior:

  • indexed filters narrow the candidate set for thought_type, role, agent_id, tags, and concepts
  • text is a case-insensitive substring match over:
    • thought content
    • agent_id
    • tags
    • concepts
    • agent-registry display name, aliases, owner, and description
  • results are returned in append order
  • limit keeps the newest matching tail after filtering rather than applying a ranking score

That means plain ThoughtQuery / /v1/search behavior is deterministic and explainable, but that baseline path is not BM25, hybrid, or vector retrieval. Additive ranked and graph-aware retrieval now exist on separate crate, REST, and MCP surfaces.

Examples:

use mentisdb::{MentisDb, ThoughtQuery, ThoughtType};
use std::path::PathBuf;

# fn main() -> std::io::Result<()> {
let chain = MentisDb::open(&PathBuf::from("/tmp/tc_query"), "agent1", "Agent", None, None)?;

let lexical = ThoughtQuery::new()
    .with_types(vec![ThoughtType::Decision])
    .with_tags_any(["search"])
    .with_text("latency");

let results = chain.query(&lexical);
# let _ = results;
# Ok(())
# }
{
  "chain_key": "mentisdb",
  "thought_types": ["Decision"],
  "tags_any": ["search"],
  "text": "latency",
  "limit": 20
}

Design note:

  • treat this lexical/filter-first behavior as the baseline
  • keep ranked, vector, and future hybrid search as separate, explicitly documented surfaces
  • do not silently change the semantics of ThoughtQuery or /v1/search from append-order filtering to score-ranked retrieval

The dedicated benchmark benches/search_baseline.rs and evaluation tests in tests/search_eval_tests.rs are intended to preserve that baseline while world-class search evolves.

Ranked Search

MentisDB now also exposes an additive ranked-search surface for direct crate use:

  • RankedSearchQuery
  • RankedSearchGraph
  • MentisDb::query_context_bundles(&RankedSearchQuery)
  • MentisDb::query_ranked(&RankedSearchQuery)
  • RankedSearchBackend::{Lexical, LexicalGraph, Heuristic}

This surface is intentionally separate from ThoughtQuery.

ThoughtQuery still decides which thoughts are eligible. Ranked search then decides how those eligible thoughts are ordered.

Current ranked-search behavior:

  • RankedSearchQuery.filter uses the same deterministic semantics as MentisDb::query
  • when text normalizes to a non-empty query, the backend is lexical
  • lexical ranking scores indexed thought text plus agent metadata from the filtered candidate set
  • when graph is enabled alongside non-empty text, the backend becomes lexical_graph
  • graph expansion starts from lexical seed hits, walks refs and typed relations, and can surface supporting context that did not lexically match
  • when text is absent or blank, the backend falls back to heuristic
  • heuristic ordering uses lightweight importance, confidence, and recency signals
  • total_candidates counts the hits after filter application and lexical gating, before final limit truncation
  • each ranked hit includes matched_terms plus match_sources such as content, tags, concepts, agent_id, and agent_registry
  • graph-expanded hits also expose graph_distance, graph_seed_paths, graph_relation_kinds, and graph_path provenance so callers can explain why a supporting thought surfaced
  • grouped context delivery is available through query_context_bundles, which anchors supporting graph hits beneath lexical seeds in deterministic order

Example:

use mentisdb::{MentisDb, RankedSearchGraph, RankedSearchQuery, ThoughtQuery, ThoughtType};
use mentisdb::search::GraphExpansionMode;
use std::path::PathBuf;

# fn main() -> std::io::Result<()> {
let chain = MentisDb::open(&PathBuf::from("/tmp/tc_ranked"), "agent1", "Agent", None, None)?;

let ranked = RankedSearchQuery::new()
    .with_filter(
        ThoughtQuery::new()
            .with_types(vec![ThoughtType::Decision])
            .with_tags_any(["search"]),
    )
    .with_text("latency ranking")
    .with_graph(
        RankedSearchGraph::new()
            .with_max_depth(1)
            .with_mode(GraphExpansionMode::Bidirectional),
    )
    .with_limit(10);

let results = chain.query_ranked(&ranked);
# let _ = results;
# Ok(())
# }

Grouped context example:

use mentisdb::{MentisDb, RankedSearchGraph, RankedSearchQuery, ThoughtQuery};
use mentisdb::search::GraphExpansionMode;
use std::path::PathBuf;

# fn main() -> std::io::Result<()> {
let chain = MentisDb::open(&PathBuf::from("/tmp/tc_ranked"), "agent1", "Agent", None, None)?;

let bundles = chain.query_context_bundles(
    &RankedSearchQuery::new()
        .with_filter(ThoughtQuery::new().with_tags_any(["search"]))
        .with_text("latency ranking")
        .with_graph(
            RankedSearchGraph::new()
                .with_mode(GraphExpansionMode::Bidirectional)
                .with_max_depth(2),
        )
        .with_limit(5),
);
# let _ = bundles;
# Ok(())
# }

Product rule:

  • keep ThoughtQuery stable and explainable for append-order filtering
  • evolve ranked search as a separate surface with its own benchmarks, tests, and transport layers
  • treat registry-aware filtering and future transport exposure as additive work on top of the current crate API
  • use query_ranked for flat ranked retrieval and query_context_bundles when the caller wants seed-anchored support context instead of one mixed list

The ranked-search benchmark benches/search_ranked.rs and evaluation tests in tests/search_ranked_eval_tests.rs are the guardrails for that additive surface.

Vector Sidecars

MentisDB now exposes an additive Phase 3 vector sidecar surface for direct crate use:

  • search::EmbeddingProvider
  • search::EmbeddingMetadata
  • search::VectorSidecar
  • VectorSearchQuery
  • MentisDb::vector_sidecar_path(&EmbeddingMetadata)
  • MentisDb::load_vector_sidecar(&EmbeddingMetadata)
  • MentisDb::vector_sidecar_freshness(&VectorSidecar, &EmbeddingMetadata)
  • MentisDb::rebuild_vector_sidecar(&provider)
  • MentisDb::query_vector(&provider, &VectorSearchQuery)

Contract:

  • embeddings remain optional, and MentisDB still works with no vector dependencies at all
  • vector state lives in a rebuildable sidecar, never in the canonical append-only chain
  • vector sidecars are separated by chain_key, thought_id, thought_hash, model_id, embedding dimension, and embedding version
  • changing the embedding model or version invalidates old vector state instead of silently mixing incompatible embeddings
  • vector hits surface whether they came from a Fresh or stale sidecar
  • deleting or corrupting the sidecar degrades only vector retrieval; plain chain reads, appends, and lexical/graph search still work

Operational flow:

  • rebuild a sidecar explicitly for one provider and chain
  • load or query that sidecar later with the same embedding metadata
  • if the chain head changes, the sidecar becomes stale and results report that freshness state until the sidecar is rebuilt

REST Lexical Search

The daemon also exposes the Phase 1 ranked lexical surface over REST at POST /v1/lexical-search.

Request shape:

{
  "chain_key": "mentisdb",
  "text": "latency ranking",
  "agent_ids": ["planner"],
  "thought_types": ["Decision"],
  "offset": 0,
  "limit": 10
}

Phase 4 Transport Contract (Ranked + Bundles)

Phase 4 transport work keeps plain POST /v1/search and POST /v1/lexical-search compatibility and adds two additive endpoints:

  • POST /v1/ranked-search for flat ranked retrieval
  • POST /v1/context-bundles for seed-anchored grouped support context

Ranked response contract fields:

  • backend
  • results[].score.{lexical,graph,relation,seed_support,importance,confidence,recency,total}
  • results[].matched_terms
  • results[].match_sources
  • results[].graph_distance
  • results[].graph_seed_paths
  • results[].graph_relation_kinds
  • results[].graph_path

Context-bundle response contract fields:

  • total_bundles
  • consumed_hits
  • bundles[].seed.{locator,lexical_score,matched_terms,thought}
  • bundles[].support[].{locator,thought,depth,seed_path_count,relation_kinds,path}

MCP transport mirrors this split with additive tools:

  • mentisdb_ranked_search
  • mentisdb_context_bundles

Acceptance coverage for these transport contracts lives in:

  • tests/search_transport_contract_tests.rs

Response shape:

{
  "total": 2,
  "results": [
    {
      "thought": { "index": 42, "agent_id": "planner", "content": "..." },
      "score": {
        "lexical": 2.91,
        "graph": 0.18,
        "relation": 0.05,
        "seed_support": 0.0,
        "importance": 0.0,
        "confidence": 0.0,
        "recency": 0.0,
        "total": 3.14
      },
      "matched_terms": ["latency", "ranking"],
      "match_sources": ["content", "tags", "agent_registry"]
    }
  ]
}

Web Dashboard

The daemon includes an embedded browser UI at:

https://127.0.0.1:9475/dashboard

The dashboard is served over HTTPS with the same self-signed certificate used by the HTTPS MCP and REST surfaces.

Dashboard capabilities:

  • live chain listing with thought and agent counts
  • thought exploration with grouped ThoughtType filters, refs, and typed relations
  • chain-scoped ranked search with text and live-agent filters
  • grouped context bundles for seed-anchored supporting search context
  • ranked result inspection in the thought modal, including score breakdowns, matched terms, graph distance, relation kinds, and bundle support preview
  • agent detail management for display name, description, owner, status, and signing keys
  • latest agent-thought browsing without restarting the daemon after new thoughts are appended
  • chain import from MEMORY.md
  • cross-chain agent-memory copy with agent metadata preserved on the target chain
  • skill browsing, diffing, deprecation, and revocation

Protect the dashboard with MENTISDB_DASHBOARD_PIN whenever the daemon is reachable outside localhost.


MCP Tool Catalog

The daemon currently exposes 33 MCP tools:

  • mentisdb_bootstrap Create a chain if needed and write one bootstrap checkpoint when it is empty.
  • mentisdb_append Append a durable semantic thought with optional tags, concepts, refs, and signature metadata.
  • mentisdb_append_retrospective Append a retrospective memory intended to prevent future agents from repeating a hard failure.
  • mentisdb_search Search thoughts by semantic filters, identity filters, time bounds, and scoring thresholds.
  • mentisdb_lexical_search Return flat ranked lexical matches with explainable term and field provenance.
  • mentisdb_ranked_search Return flat ranked lexical, graph-aware, or heuristic results with additive score breakdowns.
  • mentisdb_context_bundles Return seed-anchored grouped support context beneath the best lexical seeds.
  • mentisdb_list_chains List known chains with version, storage adapter, counts, and storage location.
  • mentisdb_list_agents List the distinct agent identities participating in one chain.
  • mentisdb_get_agent Return one full agent registry record, including status, aliases, description, keys, and per-chain activity metadata.
  • mentisdb_list_agent_registry Return the full per-chain agent registry.
  • mentisdb_upsert_agent Create or update a registry record before or after an agent writes thoughts.
  • mentisdb_set_agent_description Set or clear the description stored for one registered agent.
  • mentisdb_add_agent_alias Add a historical or alternate alias to a registered agent.
  • mentisdb_add_agent_key Add or replace one public verification key on a registered agent.
  • mentisdb_revoke_agent_key Revoke one previously registered public key.
  • mentisdb_disable_agent Disable one agent by marking its registry status as revoked.
  • mentisdb_recent_context Render recent thoughts into a prompt snippet for session resumption.
  • mentisdb_memory_markdown Export a MEMORY.md-style Markdown view of the full chain or a filtered subset.
  • mentisdb_import_memory_markdown Import a MEMORY.md-formatted Markdown document into a target chain.
  • mentisdb_get_thought Return one stored thought by stable id, chain index, or content hash.
  • mentisdb_get_genesis_thought Return the first thought ever recorded in the chain, if any.
  • mentisdb_traverse_thoughts Traverse the chain forward or backward in append order from a chosen anchor, in chunks, with optional filters.
  • mentisdb_skill_md Return the official embedded MENTISDB_SKILL.md Markdown file.
  • mentisdb_list_skills List versioned skill summaries from the skill registry.
  • mentisdb_skill_manifest Return the versioned skill-registry manifest, including searchable fields and supported formats.
  • mentisdb_upload_skill Upload a new immutable skill version from Markdown or JSON.
  • mentisdb_search_skill Search skills by indexed metadata such as ids, names, tags, triggers, uploader identity, status, format, schema version, and time window.
  • mentisdb_read_skill Read one stored skill as Markdown or JSON. Responses include trust warnings for untrusted or malicious skill content.
  • mentisdb_skill_versions List immutable uploaded versions for one skill.
  • mentisdb_deprecate_skill Mark a skill as deprecated while preserving all prior versions.
  • mentisdb_revoke_skill Mark a skill as revoked while preserving audit history.
  • mentisdb_head Return head metadata, the latest thought at the current chain tip, and integrity state.

The detailed request and response shapes for the MCP surface live in MENTISDB_MCP.md. The REST equivalents live in MENTISDB_REST.md.


Thought Lookup And Traversal

MentisDB distinguishes three different read patterns:

  • head means the newest thought at the current tip of the append-only chain
  • genesis means the very first thought in the chain
  • traversal means sequential browsing by append order, forward or backward, in chunks

That traversal model is deliberately different from graph/context traversal through refs and typed relations. Graph traversal answers "what is connected to this thought?" Sequential traversal answers "what came before or after this thought in the ledger?"

Lookup and traversal support:

  • direct thought lookup by id, hash, or index
  • logical genesis and head anchors
  • forward and backward traversal directions
  • include_anchor control for inclusive vs exclusive paging
  • chunked pagination, including chunk_size = 1 for next/previous behavior
  • optional filters reused from thought search, such as agent identity, thought type, role, tags, concepts, text, importance, confidence, and time windows
  • numeric time windows expressed as start + delta with seconds or milliseconds units for MCP/REST callers

Skill Registry

MentisDB includes a versioned skill registry stored alongside chain data in a binary file. Skills are ingested through adapters:

  • Markdown -> SkillDocument
  • JSON -> SkillDocument
  • SkillDocument -> Markdown
  • SkillDocument -> JSON

Each uploaded skill version records:

  • registry file version
  • skill schema version
  • upload timestamp
  • responsible agent_id
  • optional agent display name and owner from the MentisDB agent registry
  • source format
  • integrity hash

Uploaders must already exist in the agent registry for the referenced chain. Reusing an existing skill_id creates a new immutable version; it does not overwrite history.

read_skill responses include explicit safety warnings because SKILL.md content can be malicious. Treat every skill as advisory until provenance, trust, and requested capabilities are validated.

Skill Versioning

Each upload to an existing skill_id creates a new immutable version rather than overwriting history:

  • The first upload stores the full content (SkillVersionContent::Full).
  • Subsequent uploads store a unified diff patch against the previous version (SkillVersionContent::Delta), keeping storage efficient for iteratively improved skills.
  • Each version receives a monotone version_number (0-based, assigned in append order).
  • Pass a version_id to read_skill / mentisdb_read_skill to retrieve any historical version. The system reconstructs it by replaying patches forward from version 0.
  • skill_versions / mentisdb_skill_versions lists all versions with their ids, numbers, and timestamps.

Signed Skill Uploads

Agents that have registered Ed25519 public keys in the agent registry must sign their uploads.

Required fields when the uploading agent has active keys:

  • signing_key_id — the key_id registered via POST /v1/agents/keys or mentisdb_add_agent_key
  • skill_signature — 64-byte Ed25519 signature over the raw skill content bytes

Agents without registered public keys may upload without signatures.

Upload flow for signing agents:

  1. Register a public key:
    POST /v1/agents/keys   { agent_id, key_id, algorithm: "ed25519", public_key_bytes }
    
    or via MCP: mentisdb_add_agent_key
  2. Sign the raw content bytes with the corresponding private key (Ed25519).
  3. Include signing_key_id and skill_signature in the upload request:
    POST /v1/skills/upload   { agent_id, skill_id, content, signing_key_id, skill_signature }
    
    or via MCP: mentisdb_upload_skill with the same fields.

Using With MCP Clients

mentisdbd exposes both:

  • a standard streamable HTTP MCP endpoint at POST /
  • the legacy CloudLLM-compatible MCP endpoints at POST /tools/list and POST /tools/execute

That means you can:

  • use native MCP clients such as Codex and Claude Code against http://127.0.0.1:9471
  • keep using direct HTTP calls or cloudllm's MCP compatibility layer when needed

Codex

Codex CLI expects a streamable HTTP MCP server when you use --url:

codex mcp add mentisdb --url http://127.0.0.1:9471

Useful follow-up commands:

codex mcp list
codex mcp get mentisdb

This connects Codex to the daemon's standard MCP root endpoint.

Qwen Code

Qwen Code uses the same HTTP MCP transport model:

qwen mcp add --transport http mentisdb http://127.0.0.1:9471

Useful follow-up commands:

qwen mcp list

For user-scoped configuration:

qwen mcp add --scope user --transport http mentisdb http://127.0.0.1:9471

Claude for Desktop

Claude for Desktop connects to MCP servers through claude_desktop_config.json. It requires mcp-remote as a bridge between the desktop app and the MentisDB HTTPS endpoint.

Step 1 — Install mcp-remote (Node.js required):

npm install -g mcp-remote

Step 2 — Edit the config file (location by OS):

OS Path
macOS ~/Library/Application Support/Claude/claude_desktop_config.json
Windows %APPDATA%\Claude\claude_desktop_config.json
Linux ~/.config/Claude/claude_desktop_config.json

macOS:

{
  "mcpServers": {
    "mentisdb": {
      "command": "/opt/homebrew/bin/mcp-remote",
      "args": ["https://my.mentisdb.com:9473"],
      "env": { "NODE_TLS_REJECT_UNAUTHORIZED": "0" }
    }
  }
}

Windows:

{
  "mcpServers": {
    "mentisdb": {
      "command": "mcp-remote",
      "args": ["https://my.mentisdb.com:9473"],
      "env": { "NODE_TLS_REJECT_UNAUTHORIZED": "0" }
    }
  }
}

If Windows can't find the binary, supply the full path: C:\Users\YourName\AppData\Roaming\npm\mcp-remote.cmd

Linux:

{
  "mcpServers": {
    "mentisdb": {
      "command": "/usr/local/bin/mcp-remote",
      "args": ["https://my.mentisdb.com:9473"],
      "env": { "NODE_TLS_REJECT_UNAUTHORIZED": "0" }
    }
  }
}

Use which mcp-remote to confirm the binary path on your machine.

Why NODE_TLS_REJECT_UNAUTHORIZED: "0"?
MentisDB ships with a self-signed TLS certificate. Node.js rejects self-signed certs by default, which causes mcp-remote to disconnect immediately after the MCP initialize handshake. This env var disables that check for the mcp-remote process only. As an alternative, trust the certificate at the OS level (sudo security add-trusted-cert on macOS) and remove the env block.

Restart Claude for Desktop after saving the config file.

Claude Code

Claude Code supports MCP servers through its claude mcp commands and project/user MCP config. For a remote HTTP MCP server, the configuration shape is transport-based:

claude mcp add --transport http mentisdb http://127.0.0.1:9471

Useful follow-up commands:

claude mcp list
claude mcp get mentisdb

mentisdbd setup claude-code merges the MCP server entry into ~/.claude/settings.json, preserving your existing Claude Code settings. The MentisDB HTTP MCP block it writes looks like this:

{
  "mcpServers": {
    "mentisdb": {
      "type": "http",
      "url": "http://127.0.0.1:9471"
    }
  }
}

Important:

  • /mcp inside Claude Code is mainly for managing or authenticating MCP servers that are already configured
  • the server itself must already be running at the configured URL

GitHub Copilot CLI

GitHub Copilot CLI can also connect to mentisdbd as a remote HTTP MCP server.

From interactive mode:

  1. Run /mcp add
  2. Set Server Name to mentisdb
  3. Set Server Type to HTTP
  4. Set URL to http://127.0.0.1:9471
  5. Leave headers empty unless you add auth later
  6. Save the config

You can also configure it manually in ~/.copilot/mcp-config.json (or $XDG_CONFIG_HOME/copilot/mcp-config.json when XDG_CONFIG_HOME is set):

{
  "mcpServers": {
    "mentisdb": {
      "type": "http",
      "url": "http://127.0.0.1:9471",
      "headers": {},
      "tools": ["*"]
    }
  }
}

Retrospective Memory

MentisDB supports a dedicated retrospective workflow for lessons learned.

  • Use mentisdb_append for ordinary durable facts, constraints, decisions, plans, and summaries.
  • Use mentisdb_append_retrospective after a repeated failure, a long snag, or a non-obvious fix when future agents should avoid repeating the same struggle.

The retrospective helper:

  • defaults thought_type to LessonLearned
  • always stores the thought with role = Retrospective
  • still supports tags, concepts, confidence, importance, and refs to earlier thoughts such as the original mistake or correction

Thought Types And Roles

MentisDB currently defines 29 semantic ThoughtType values and 8 operational ThoughtRole values.

Thought types:

  • PreferenceUpdate, UserTrait, RelationshipUpdate
  • Finding, Insight, FactLearned, PatternDetected, Hypothesis, Surprise
  • Mistake, Correction, LessonLearned, AssumptionInvalidated, Reframe
  • Constraint, Plan, Subgoal, Decision, StrategyShift
  • Wonder, Question, Idea, Experiment
  • ActionTaken, TaskComplete
  • Checkpoint, StateSnapshot, Handoff, Summary

Thought roles:

  • Memory
  • WorkingMemory
  • Summary
  • Compression
  • Checkpoint
  • Handoff
  • Audit
  • Retrospective

Use ThoughtType to say what the memory means semantically, and ThoughtRole to say how the system should treat it operationally. The crate rustdoc is the authoritative source for per-variant semantics, and the Agent Guide on the docs site contains a human-oriented explanation of when to use each one.


Shared-Chain Multi-Agent Use

Multiple agents can write to the same chain_key.

Each stored thought carries a stable:

  • agent_id

Agent profile metadata now lives in the per-chain agent registry instead of being duplicated into every thought record. Registry records can store:

  • display_name
  • agent_owner
  • description
  • aliases
  • status
  • public_keys
  • per-chain activity counters such as thought_count, first_seen_index, and last_seen_index

That allows a shared chain to represent memory from:

  • multiple agents in one workflow
  • multiple named roles in one orchestration system
  • multiple tenants or owners writing to the same chain namespace

Queries can filter by:

  • agent_id
  • agent_name
  • agent_owner

Administrative tools can also inspect and mutate the agent registry directly, so agents can be documented, disabled, aliased, or provisioned with public keys before they start writing thoughts.


Related Docs

At the repository root:

  • MENTISDB_MCP.md
  • MENTISDB_REST.md
  • mentisdb/WHITEPAPER.md
  • mentisdb/changelog.txt