m1nd-core 0.4.0

Neuro-symbolic code graph engine — core algorithms, graph, plasticity, resonance
Documentation

🇬🇧 English | 🇧🇷 Português | 🇪🇸 Español | 🇮🇹 Italiano | 🇫🇷 Français | 🇩🇪 Deutsch | 🇨🇳 中文


m1nd doesn't search your codebase -- it activates it. Fire a query into the graph and watch signal propagate across structural, semantic, temporal, and causal dimensions. Noise cancels out. Relevant connections amplify. And the graph learns from every interaction via Hebbian plasticity.

335 files -> 9,767 nodes -> 26,557 edges in 0.91 seconds.
Then: activate in 31ms. impact in 5ms. trace in 3.5ms. learn in <1ms.

Proven Results

Live audit on a production Python/FastAPI codebase (52K lines, 380 files):

Metric Result
Bugs found in one session 39 (28 confirmed fixed + 9 high-confidence)
Invisible to grep 8 of 28 (28.5%) -- required structural analysis
Hypothesis accuracy 89% over 10 live claims
LLM tokens consumed 0 -- pure Rust, local binary
m1nd queries vs grep ops 46 vs ~210
Total query latency ~3.1 seconds vs ~35 minutes estimated

Criterion micro-benchmarks (real hardware):

Operation Time
activate 1K nodes 1.36 µs
impact depth=3 543 ns
flow_simulate 4 particles 552 µs
antibody_scan 50 patterns 2.68 ms
layer_detect 500 nodes 862 µs
resonate 5 harmonics 8.17 µs

Quick Start

git clone https://github.com/maxkle1nz/m1nd.git
cd m1nd && cargo build --release
./target/release/m1nd-mcp
// 1. Ingest your codebase (910ms for 335 files)
{"method":"tools/call","params":{"name":"m1nd.ingest","arguments":{"path":"/your/project","agent_id":"dev"}}}
// -> 9,767 nodes, 26,557 edges, PageRank computed

// 2. Ask: "What's related to authentication?"
{"method":"tools/call","params":{"name":"m1nd.activate","arguments":{"query":"authentication","agent_id":"dev"}}}
// -> auth fires -> propagates to session, middleware, JWT, user model
//    ghost edges reveal undocumented connections

// 3. Tell the graph what was useful
{"method":"tools/call","params":{"name":"m1nd.learn","arguments":{"feedback":"correct","node_ids":["file::auth.py","file::middleware.py"],"agent_id":"dev"}}}
// -> 740 edges strengthened via Hebbian LTP. Next query is smarter.

Add to Claude Code (~/.claude.json):

{
  "mcpServers": {
    "m1nd": {
      "command": "/path/to/m1nd-mcp",
      "env": {
        "M1ND_GRAPH_SOURCE": "/tmp/m1nd-graph.json",
        "M1ND_PLASTICITY_STATE": "/tmp/m1nd-plasticity.json"
      }
    }
  }
}

Works with any MCP client: Claude Code, Cursor, Windsurf, Zed, or your own.


It worked? Star this repo -- it helps others find it. Bug or idea? Open an issue. Want to go deeper? See EXAMPLES.md for real-world pipelines.


Why Not Just Use Cursor/RAG/grep?

Capability Sourcegraph Cursor Aider RAG m1nd
Code graph SCIP (static) Embeddings tree-sitter + PageRank None CSR + 4D activation
Learns from use No No No No Hebbian plasticity
Persists investigations No No No No Trail save/resume/merge
Tests hypotheses No No No No Bayesian on graph paths
Simulates removal No No No No Counterfactual cascade
Multi-repo graph Search only No No No Federated graph
Temporal intelligence git blame No No No Co-change + velocity + decay
Ingests docs + code No No No Partial Memory adapter (typed graph)
Bug immune memory No No No No Antibody system
Pre-failure detection No No No No Tremor + epidemic + trust
Architectural layers No No No No Auto-detect + violation report
Cost per query Hosted SaaS Subscription LLM tokens LLM tokens Zero

Comparisons reflect capabilities at time of writing. Each tool excels at its primary use case; m1nd is not a replacement for Sourcegraph's enterprise search or Cursor's editing UX.

What Makes It Different

The graph learns. Confirm results are useful -- edge weights strengthen (Hebbian LTP). Mark results wrong -- they weaken (LTD). The graph evolves to match how your team thinks about your codebase. No other code intelligence tool does this.

The graph tests claims. "Does worker_pool depend on whatsapp_manager at runtime?" m1nd explores 25,015 paths in 58ms and returns a Bayesian confidence verdict. 89% accuracy across 10 live claims. It confirmed a session_pool leak at 99% confidence (3 bugs found) and correctly rejected a circular dependency hypothesis at 1%.

The graph ingests memory. Pass adapter: "memory" to ingest .md/.txt files into the same graph as code. activate("antibody pattern matching") returns both pattern_models.py (implementation) and PRD-ANTIBODIES.md (spec). missing("GUI web server") finds specs with no implementation -- gap detection across domains.

The graph detects bugs before they happen. Five engines beyond structural analysis:

  • Antibody System -- remembers bug patterns, scans for recurrence on every ingest
  • Epidemic Engine -- SIR propagation predicts which modules harbor undiscovered bugs
  • Tremor Detection -- change acceleration (second derivative) precedes bugs, not just churn
  • Trust Ledger -- per-module actuarial risk scores from defect history
  • Layer Detection -- auto-detects architectural layers, reports dependency violations

The graph saves investigations. trail.save -> trail.resume days later from the exact same cognitive position. Two agents on the same bug? trail.merge -- automatic conflict detection on shared nodes.

The 61 Tools

Category Count Highlights
Foundation 13 ingest, activate, impact, why, learn, drift, seek, scan, warmup, federate
Perspective Navigation 12 Navigate the graph like a filesystem -- start, follow, peek, branch, compare
Lock System 5 Pin subgraph regions, watch for changes (lock.diff: 0.08µs)
Superpowers 13 hypothesize, counterfactual, missing, resonate, fingerprint, trace, predict, trails
Superpowers Extended 9 antibody, flow_simulate, epidemic, tremor, trust, layers
Surgical 4 surgical_context, apply, surgical_context_v2, apply_batch
Intelligence 5 search, help, panoramic, savings, report
Tool What It Does Speed
ingest Parse codebase into semantic graph 910ms / 335 files
activate Spreading activation with 4D scoring 1.36µs (bench)
impact Blast radius of a code change 543ns (bench)
why Shortest path between two nodes 5-6ms
learn Hebbian feedback -- graph gets smarter <1ms
drift What changed since last session 23ms
health Server diagnostics <1ms
seek Find code by natural language intent 10-15ms
scan 8 structural patterns (concurrency, auth, errors...) 3-5ms each
timeline Temporal evolution of a node ~ms
diverge Structural divergence analysis varies
warmup Prime graph for an upcoming task 82-89ms
federate Unify multiple repos into one graph 1.3s / 2 repos
Tool Purpose
perspective.start Open a perspective anchored to a node
perspective.routes List available routes from current focus
perspective.follow Move focus to a route target
perspective.back Navigate backward
perspective.peek Read source code at the focused node
perspective.inspect Deep metadata + 5-factor score breakdown
perspective.suggest Navigation recommendation
perspective.affinity Check route relevance to current investigation
perspective.branch Fork an independent perspective copy
perspective.compare Diff two perspectives (shared/unique nodes)
perspective.list All active perspectives + memory usage
perspective.close Release perspective state
Tool Purpose Speed
lock.create Snapshot a subgraph region 24ms
lock.watch Register change strategy ~0ms
lock.diff Compare current vs baseline 0.08µs
lock.rebase Advance baseline to current 22ms
lock.release Free lock state ~0ms
Tool What It Does Speed
hypothesize Test claims against graph structure (89% accuracy) 28-58ms
counterfactual Simulate module removal -- full cascade 3ms
missing Find structural holes 44-67ms
resonate Standing wave analysis -- find structural hubs 37-52ms
fingerprint Find structural twins by topology 1-107ms
trace Map stacktraces to root causes 3.5-5.8ms
validate_plan Pre-flight risk assessment for changes 0.5-10ms
predict Co-change prediction <1ms
trail.save Persist investigation state ~0ms
trail.resume Restore exact investigation context 0.2ms
trail.merge Combine multi-agent investigations 1.2ms
trail.list Browse saved investigations ~0ms
differential Structural diff between graph snapshots ~ms
Tool What It Does Speed
antibody_scan Scan graph against stored bug patterns 2.68ms
antibody_list List stored antibodies with match history ~0ms
antibody_create Create, disable, enable, or delete an antibody ~0ms
flow_simulate Concurrent execution flow -- race condition detection 552µs
epidemic SIR bug propagation prediction 110µs
tremor Change frequency acceleration detection 236µs
trust Per-module defect history trust scores 70µs
layers Auto-detect architectural layers + violations 862µs
layer_inspect Inspect a specific layer: nodes, edges, health varies
Tool What It Does Speed
surgical_context Complete context for a code node: source, callers, callees, tests, trust score, blast radius — in one call varies
apply Write edited code back to file, atomic write, re-ingest graph, run predict 3.5ms
surgical_context_v2 All connected files with source code in ONE call — complete dependency context without multiple round-trips 1.3ms
apply_batch Write multiple files atomically, single re-ingest pass, returns per-file diffs 165ms
Tool What It Does Speed
search Literal + regex full-text search across all graph node labels and source content 4-11ms
help Built-in tool reference — per-tool docs, parameters, and usage examples <1ms
panoramic Full-codebase risk panorama — 50 modules scanned, risk scores ranked 38ms
savings Token economy tracker — cumulative LLM tokens saved vs. direct-read baseline <1ms
report Structured session report — metrics, top nodes, anomalies, savings in markdown <1ms

Full API reference with examples ->

Architecture

Three Rust crates. No runtime dependencies. No LLM calls. No API keys. ~8MB binary.

m1nd-core/     Graph engine, spreading activation, Hebbian plasticity, hypothesis engine,
               antibody system, flow simulator, epidemic, tremor, trust, layer detection
m1nd-ingest/   Language extractors (28 languages), memory adapter, JSON adapter,
               git enrichment, cross-file resolver, incremental diff
m1nd-mcp/      MCP server, 61 tool handlers, JSON-RPC over stdio, HTTP server + GUI
graph LR
    subgraph Ingest
        A[Code / 28 langs] --> R[Reference Resolver]
        MA[Memory adapter] --> R
        JA[JSON adapter] --> R
        R --> GD[Git enrichment]
        GD --> G[CSR Graph]
    end
    subgraph Core
        G --> SA[Spreading Activation]
        G --> HP[Hebbian Plasticity]
        G --> HY[Hypothesis Engine]
        G --> SX[Superpowers Extended]
        SA --> XLR[XLR Noise Cancel]
    end
    subgraph MCP
        XLR --> T[61 Tools]
        HP --> T
        HY --> T
        SX --> T
        T --> IO[JSON-RPC stdio]
        T --> HTTP[HTTP API + GUI]
    end
    IO --> C[Claude Code / Cursor / any MCP client]
    HTTP --> B[Browser on localhost:1337]

28 languages via tree-sitter across two tiers. Default build includes Tier 2 (8 langs). Add --features tier1 for all 28. Language details ->

When NOT to Use m1nd

  • You need neural semantic search. V1 uses trigram matching, not embeddings. "Find code that means authentication but never uses the word" won't work yet.
  • You have 400K+ files. The graph lives in memory (~2MB per 10K nodes). It works, but it wasn't optimized for that scale.
  • You need dataflow / taint analysis. m1nd tracks structural and co-change relationships, not data propagation through variables. Use Semgrep or CodeQL for that.
  • You need sub-symbol tracking. m1nd models function calls and imports as edges, not data flow through arguments.
  • You need real-time indexing on every save. Ingest is fast (910ms for 335 files) but not instantaneous. m1nd is for session-level intelligence, not keystroke feedback. Use your LSP for that.

Use Cases

Bug hunt: hypothesize -> missing -> flow_simulate -> trace. Zero grep. The graph navigates to the bug. 39 bugs found in one session.

Pre-deploy gate: antibody_scan -> validate_plan -> epidemic. Scans for known bug shapes, assesses blast radius, predicts infection spread.

Architecture audit: layers -> layer_inspect -> counterfactual. Auto-detects layers, finds violations, simulates what breaks if you remove a module.

Onboarding: activate -> layers -> perspective.start -> perspective.follow. New developer asks "how does auth work?" -- graph lights up the path.

Cross-domain search: ingest(adapter="memory", mode="merge") -> activate. Code + docs in one graph. One question returns both the spec and the implementation.

Contributing

m1nd is early-stage and evolving fast. Contributions welcome: language extractors, graph algorithms, MCP tools, and benchmarks. See CONTRIBUTING.md.

License

MIT -- see LICENSE.