Skip to main content
Glama
star-ga
by star-ga

Drop-in memory layer for AI coding agents — Claude Code, Claude Desktop, Codex CLI, Gemini CLI, Cursor, Windsurf, Zed, OpenClaw, or any MCP-compatible client. Upgrades your agent from "chat history + notes" to a governed Memory OS with hybrid search, RRF fusion, intent routing, optional MIND kernels, structured persistence, contradiction detection, drift analysis, safe governance, and full audit trail.

If your agent runs for weeks, it will drift. mind-mem prevents silent drift.

mind-mem powers the Memory Plane of the MIND Cognitive Kernel — the deterministic AI runtime architecture.

Shared Memory Across All Your AI Agents

This is the killer feature. When you install mind-mem, all your AI coding agents share the same memory workspace. Claude Code, Codex CLI, Gemini CLI, Cursor, Windsurf, Zed — every MCP-compatible client connects to the same persistent memory through a single workspace.

What this means in practice:

  • A decision made in Claude Code is instantly recalled by Codex CLI and Gemini CLI

  • Entity knowledge (projects, tools, people) accumulates from all sessions across all agents

  • Contradictions detected by one agent are flagged to all others

  • Your memory doesn't fragment across tools — it compounds

One install script, all agents configured in seconds:

git clone https://github.com/star-ga/mind-mem.git
cd mind-mem
./install.sh --all    # Auto-detects and configures every AI coding client on your machine

The installer auto-detects Claude Code, Claude Desktop, Codex CLI, Gemini CLI, Cursor, Windsurf, Zed, and OpenClaw — creates a shared workspace and wires the MCP server into each client's config. SQLite WAL mode ensures safe concurrent access: one writer, many readers, zero corruption.

30-Second Demo

pip install mind-mem
mind-mem-init ~/my-workspace        # Create workspace
mind-mem-recall -q "API decisions" --workspace ~/my-workspace  # Hybrid BM25F search
mind-mem-scan ~/my-workspace        # Detect drift & contradictions

Output:

[1.204] D-20260215-001 (decision) — Use async/await for all API endpoints
        decisions/DECISIONS.md:11
[1.094] D-20260210-003 (decision) — REST over GraphQL for public API
        decisions/DECISIONS.md:20

Trust Signals

Principle

What it means

Deterministic

Same input, same output. No ML in the core, no probabilistic mutations.

Auditable

Every apply logged with timestamp, receipt, and DIFF. Full traceability.

Local-first

All data stays on disk. No cloud calls, no telemetry, no phoning home.

No vendor lock-in

Plain Markdown files. Move to any system, any time.

Zero magic

Every check is a grep, every mutation is a file write. Read the source in 30 min.

No silent mutation

Nothing writes to source of truth without explicit /apply. Ever.

Zero infrastructure

No Redis, no Postgres, no vector DB, no GPU. Python 3.10+ and stdlib only.

100% NIAH

250/250 Needle In A Haystack retrieval. Every needle, every depth, every size.


Table of Contents


Why mind-mem

Most memory plugins store and retrieve. That's table stakes.

mind-mem also detects when your memory is wrong — contradictions between decisions, drift from informal choices never formalized, dead decisions nobody references, orphan tasks pointing at nothing — and offers a safe path to fix it.

Problem

Without mind-mem

With mind-mem

Contradicting decisions

Follows whichever seen last

Flags, links both, proposes fix

Informal chat decision

Lost after session ends

Auto-captured, proposed to formalize

Stale decision

Zombie confuses future sessions

Detected as dead, flagged

Orphan task reference

Silent breakage

Caught in integrity scan

Scattered recall quality

Single-mode search misses context

Hybrid BM25+Vector+RRF fusion finds it

Ambiguous query intent

One-size-fits-all retrieval

9-type intent router optimizes parameters

Novel Contributions

mind-mem introduces several techniques not found in existing memory systems:

Technique

What's new

Why it matters

Co-retrieval graph

PageRank-like score propagation across blocks frequently retrieved together

Surfaces structurally relevant blocks with zero lexical overlap (+2.0pp accuracy)

Fact card sub-block indexing

Atomic fact extraction → small-to-big retrieval with parent score blending

Catches fine-grained facts that full-block BM25 misses (+2.6pp accuracy)

Adaptive knee cutoff

Score-drop-based truncation instead of fixed top-K

Eliminates noise that hurts LLM judges — returns 3-15 results adaptively

Hard negative mining

Logs BM25-high / cross-encoder-low blocks as misleading, penalizes in future queries

Self-improving retrieval: precision increases over time without retraining

Deterministic abstention

Pre-LLM confidence gate using 5-signal scoring (entity, BM25, speaker, evidence, negation)

Prevents hallucinated answers to unanswerable questions — no ML required

Governance pipeline

Contradiction detection + drift analysis + safe apply with audit trail

Only memory system that detects when stored knowledge is wrong

Agent-agnostic shared memory

Single MCP workspace shared across Claude Code, Codex, Gemini, Cursor, Windsurf, Zed

Memory compounds across tools instead of fragmenting


Features

Hybrid BM25+Vector Search with RRF Fusion

Thread-parallel BM25 and vector search with Reciprocal Rank Fusion (k=60). Configurable weights per signal. Vector is optional — works with just BM25 out of the box.

RM3 Dynamic Query Expansion

Pseudo-relevance feedback using JM-smoothed language models. Expands queries with top terms from initial result set. Falls back to static synonyms for adversarial queries. Zero dependencies.

9-Type Intent Router

Classifies queries into WHY, WHEN, ENTITY, WHAT, HOW, LIST, VERIFY, COMPARE, or TRACE. Each intent type maps to optimized retrieval parameters (limits, expansion settings, graph traversal depth).

A-MEM Metadata Evolution

Auto-maintained per-block metadata: access counts, importance scores (clamped to [0.8, 1.5] reranking boost), keyword evolution, and co-occurrence tracking. Importance decays with exponential recency.

Deterministic Reranking

Four-signal reranking pipeline: negation awareness (penalizes contradicting results), date proximity (Gaussian decay), 20-category taxonomy matching, and recency boosting. No ML required.

Optional Cross-Encoder

Drop-in ms-marco-MiniLM-L-6-v2 cross-encoder (80MB). Blends 0.6 * CE + 0.4 * original score. Falls back gracefully when unavailable. Enabled via config.

MIND Kernels (Optional, Native Speed)

17 compiled MIND scoring kernels (BM25F, RRF fusion, reranking, negation penalty, date proximity, category boost, importance, entity overlap, confidence, top-k, weighted rank, category affinity, query-category relevance, category assignment). Compiles to native .so via the MIND compiler. Pure Python fallback always available — no functionality is lost without compilation.

BM25F Hybrid Recall

BM25F field-weighted scoring (k1=1.2, b=0.75) with per-field weighting (Statement: 3x, Title: 2.5x, Name: 2x, Summary: 1.5x), Porter stemming, bigram phrase matching (25% boost per hit), overlapping sentence chunking (3-sentence windows with 1-sentence overlap), domain-aware query expansion, and optional 2-hop graph-based cross-reference neighbor boosting. Zero dependencies. Fast and deterministic.

Graph-Based Recall

2-hop cross-reference neighbor boosting — when a keyword match is found, blocks that reference or are referenced by the match get boosted (1-hop: 0.3x decay, 2-hop: 0.1x decay). Surfaces related decisions, tasks, and entities that share no keywords but are structurally connected. Auto-enabled for multi-hop queries.

Vector Recall (optional)

Pluggable embedding backend — local ONNX (all-MiniLM-L6-v2, no server needed) or cloud (Pinecone). Falls back to BM25 when unavailable.

Persistent Memory

Structured, validated, append-only decisions / tasks / entities / incidents with provenance and supersede chains. Plain Markdown files — readable by humans, parseable by machines.

Immune System

Continuous integrity checking: contradictions, drift, dead decisions, orphan tasks, coverage scoring, regression detection. 74+ structural validation rules.

Safe Governance

All changes flow through graduated modes: detect_onlyproposeenforce. Apply engine with snapshot, receipt, DIFF, and automatic rollback on validation failure.

Adversarial Abstention Classifier

Deterministic pre-LLM confidence gate for adversarial/verification queries. Computes confidence from entity overlap, BM25 score, speaker coverage, evidence density, and negation asymmetry. Below threshold → forces abstention without calling the LLM, preventing hallucinated answers to unanswerable questions.

Auto-Capture with Structured Extraction

Session-end hook detects decision/task language (26 patterns with confidence classification), extracts structured metadata (subject, object, tags), and writes to SIGNALS.md only. Never touches source of truth directly. All signals go through /apply.

Concurrency Safety

Cross-platform advisory file locking (fcntl/msvcrt/atomic create) protects all concurrent write paths. Stale lock detection with PID-based cleanup. Zero dependencies.

Compaction & GC

Automated workspace maintenance: archive completed blocks, clean up old snapshots, compact resolved signals, archive daily logs into yearly files. Configurable thresholds with dry-run mode.

Observability

Structured JSON logging (via stdlib), in-process metrics counters, and timing context managers. All scripts emit machine-parseable events. Controlled via MIND_MEM_LOG_LEVEL env var.

Multi-Agent Namespaces & ACL

Workspace-level + per-agent private namespaces with JSON-based ACL. fnmatch pattern matching for agent policies. Shared fact ledger for cross-agent propagation with dedup and review gate.

Automated Conflict Resolution

Graduated resolution pipeline: timestamp priority, confidence priority, scope specificity, manual fallback. Generates supersede proposals with integrity hashes. Human veto loop — never auto-applies without review.

Write-Ahead Log (WAL) + Backup/Restore

Crash-safe writes via journal-based WAL. Full workspace backup (tar.gz), git-friendly JSONL export, selective restore with conflict detection and path traversal protection.

Transcript JSONL Capture

Scans Claude Code transcript files for user corrections, convention discoveries, bug fix insights, and architectural decisions. 16 transcript-specific patterns with role filtering and confidence classification.

MCP Server (21 tools, 8 resources)

Full Model Context Protocol server with 19 tools and 8 read-only resources. Works with Claude Code, Claude Desktop, Cursor, Windsurf, and any MCP-compatible client. HTTP and stdio transports with optional bearer token auth.

74+ Structural Checks + 2189 Unit Tests

validate.sh checks schemas, cross-references, ID formats, status values, supersede chains, ConstraintSignatures, and more. Backed by 2189 pytest unit tests covering all core modules.

Audit Trail

Every applied proposal logged with timestamp, receipt, and DIFF. Full traceability from signal → proposal → decision.

Calibration Feedback Loop

Per-block quality tracking with Bayesian weight computation. When users provide feedback (thumbs up/down) via calibration_feedback, the system maintains a rolling quality score per block over a 30-day window. Bayesian smoothing constrains calibration weights to the 0.5-1.5 range, preventing any single block from dominating or being silenced. Calibration weights integrate directly into the BM25 + FTS5 retrieval pipeline — high-quality blocks rank higher, low-quality blocks are naturally demoted. Use calibration_stats to inspect per-block quality distributions and global calibration health.

Feature Completeness Matrix

Capability

mind-mem

Mem0

Zep

Letta

LangMem

BM25 lexical search

Y

Vector semantic search

Y

Y

Y

Y

Y

Hybrid BM25+Vector+RRF

Y

Cross-encoder reranking

Y

Intent-aware routing (9 types)

Y

RM3 query expansion

Y

Co-retrieval graph (PageRank)

Y

Fact sub-block indexing

Y

Hard negative mining

Y

Adaptive knee cutoff

Y

Contradiction detection

Y

Drift analysis

Y

Governance pipeline (propose/apply)

Y

Multi-agent shared memory (MCP)

Y

Y

Zero core dependencies

Y

Local-only (no cloud required)

Y

Compiled native kernels (MIND)

Y

Backup/restore with zip-slip protection

Y


Benchmark Results

mind-mem's recall engine evaluated on standard long-term memory benchmarks using multiple configurations — from pure BM25 to full hybrid retrieval with neural reranking.

Needle In A Haystack (NIAH)

250/250 — 100% retrieval across all haystack sizes, burial depths, and needle types.

A single fact is planted at a controlled depth within a haystack of semantically diverse filler blocks. The system must retrieve the needle in its top-5 results using only a natural-language query.

Haystack Size

Depths Tested

Needles

Passed

Rate

10 blocks

0/25/50/75/100%

10

50/50

100%

50 blocks

0/25/50/75/100%

10

50/50

100%

100 blocks

0/25/50/75/100%

10

50/50

100%

250 blocks

0/25/50/75/100%

10

50/50

100%

500 blocks

0/25/50/75/100%

10

50/50

100%

Config: Hybrid BM25 + BAAI/bge-large-en-v1.5 + RRF (k=60) + sqlite-vec. Full details: benchmarks/NIAH.md

LoCoMo LLM-as-Judge

Same pipeline as Mem0 and Letta evaluations: retrieve context, generate answer with LLM, score against gold reference with judge LLM. Directly comparable methodology.

v1.0.7 — Hybrid + top_k=18 (Mistral answerer + judge, conv-0, 199 questions):

Category

N

Acc (>=50)

Mean Score

Overall

199

92.5%

76.7

Adversarial

47

97.9%

89.8

Multi-hop

37

91.9%

74.3

Open-domain

70

92.9%

72.7

Temporal

13

92.3%

76.2

Single-hop

32

84.4%

68.9

Pipeline: BM25 + Qwen3-Embedding-8B (4096d) vector search → RRF fusion (k=60) → top-18 evidence blocks → observation compression → answer → judge. A/B validated: +2.8 mean vs top_k=10 baseline.

v1.1.1 — BM25 + top_k=18 (Mistral Large answerer + judge, 10 conversations, 1986 questions):

Category

N

Acc (>=50)

Mean Score

Overall

1986

73.8%

70.5

Adversarial

446

92.4%

87.2

Single-hop

282

80.9%

68.7

Open-domain

841

71.2%

70.3

Temporal

96

66.7%

65.9

Multi-hop

321

50.5%

51.1

Pipeline: BM25 + RM3 query expansion → top-18 evidence blocks → observation compression → answer → judge. Full 10-conversation benchmark with Mistral Large as both answerer and judge.

v1.0.0 — BM25-only baseline (gpt-4o-mini answerer + judge, 10 conversations):

Category

N

Acc (>=50)

Mean Score

Overall

1986

67.3%

61.4

Open-domain

841

86.6%

78.3

Temporal

96

78.1%

65.7

Single-hop

282

68.8%

59.1

Multi-hop

321

55.5%

48.4

Adversarial

446

36.3%

39.5

Key improvements since v1.0.0: Adversarial accuracy tripled from 36.3% to 92.4% via abstention classifier + hybrid retrieval. Overall Acc≥50 improved from 67.3% to 73.8% (+6.5pp).

Competitive Landscape

System

Score

Approach

mind-mem

76.7%

Hybrid BM25 + Qwen3-8B vector + RRF fusion (local-only)

Memobase

75.8%

Specialized extraction

Letta

74.0%

Files + agent tool use

mind-mem

73.8%

BM25-only, full 10-conv (1986 questions, Mistral Large)

Mem0

68.5%

Graph + LLM extraction

mind-mem now surpasses Mem0 and Letta with local-only retrieval — no cloud calls, no graph DB, no LLM in the retrieval loop. mind-mem's unique value is governance (contradiction detection, drift analysis, audit trails) and agent-agnostic shared memory via MCP — areas these benchmarks don't measure.

Benchmark Comparison (2026-02-22)

System

LoCoMo Acc>=50

LongMemEval R@10

Infrastructure

Dependencies

mind-mem (hybrid)

76.7%

88.1%

Local-only

Zero core (optional: llama.cpp, sentence-transformers)

Memobase

75.8%

--

Cloud + GPU

embeddings + vector DB

Letta

74.0%

--

Cloud

embeddings + vector DB

mind-mem (BM25)

73.8%

88.1%

Local-only

Zero core

full-context

72.9%

--

N/A

LLM context window

Mem0

68.5%

--

Cloud (managed)

graph DB + embeddings

mind-mem surpasses Mem0 (68.5%), Letta (74.0%), and Memobase (75.8%) with zero cloud infrastructure. Full 10-conversation benchmark (1986 questions) validates this at scale.

LongMemEval (ICLR 2025, 470 questions)

Category

N

R@1

R@5

R@10

MRR

Overall

470

73.2

85.3

88.1

.784

Multi-session

121

83.5

95.9

95.9

.885

Temporal

127

76.4

91.3

92.9

.826

Knowledge update

72

80.6

88.9

91.7

.844

Single-session

56

82.1

89.3

89.3

.847

Performance (Latency & Throughput)

Measured on a 65-block workspace (typical personal workspace) with SQLite FTS5 backend:

Operation

Metric

Value

Query (FTS5 + rerank)

p50 latency

2.1 ms

Query (FTS5 + rerank)

p95 latency

4.9 ms

Query (FTS5 + rerank)

mean latency

2.6 ms

Incremental reindex

elapsed

32 ms (13 blocks indexed)

Full index build

elapsed

48 ms (65 blocks)

MCP tool overhead

stdio round-trip

< 15 ms

Memory footprint

RSS (idle MCP server)

~28 MB

Query latency scales as O(log N) with SQLite FTS5 (vs O(corpus) for scan backend). The co-retrieval graph adds < 1ms per query. Knee cutoff and fact aggregation add negligible overhead (< 0.5ms).

Run Benchmarks Yourself

# Retrieval-only (R@K metrics)
python3 benchmarks/locomo_harness.py
python3 benchmarks/longmemeval_harness.py

# LLM-as-judge (accuracy metrics, requires API key)
python3 benchmarks/locomo_judge.py --dry-run
python3 benchmarks/locomo_judge.py --answerer-model gpt-4o-mini --output results.json

# Hybrid retrieval with any model pair (BM25 + vector + cross-encoder)
python3 benchmarks/locomo_judge.py --hybrid --compress --answerer-model mistral-large-latest --judge-model mistral-large-latest --output results.json

# Selective conversations
python3 benchmarks/locomo_harness.py --conv-ids 4,7,8

Quick Start

pip install (quickest)

pip install mind-mem
mind-mem-recall "What decisions were made about the API?"
git clone https://github.com/star-ga/mind-mem.git
cd mind-mem
./install.sh --all

This auto-detects every AI coding client on your machine and configures mind-mem for all of them. Supported clients:

Client

Config Location

Format

Claude Code CLI

~/.claude/mcp.json

JSON

Claude Desktop

~/.config/Claude/claude_desktop_config.json

JSON

Codex CLI (OpenAI)

~/.codex/config.toml

TOML

Gemini CLI (Google)

~/.gemini/settings.json

JSON

Cursor

~/.cursor/mcp.json

JSON

Windsurf

~/.codeium/windsurf/mcp_config.json

JSON

Zed

~/.config/zed/settings.json

JSON

OpenClaw

~/.openclaw/hooks/mind-mem/

JS hook

Selective install:

./install.sh --claude-code --codex --gemini         # Only specific clients
./install.sh --all --workspace ~/my-project/memory  # Custom workspace path

Uninstall:

./uninstall.sh          # Remove from all clients (keeps workspace data)
./uninstall.sh --purge  # Remove everything including workspace data

Manual Setup

For manual or per-project setup:

1. Clone into your project

cd /path/to/your/project
git clone https://github.com/star-ga/mind-mem.git .mind-mem

2. Initialize workspace

python3 .mind-mem/src/mind_mem/init_workspace.py .

Creates 12 directories, 19 template files, and mind-mem.json config. Never overwrites existing files.

3. Validate

bash .mind-mem/src/mind_mem/validate.sh .
# or cross-platform:
python3 .mind-mem/src/mind_mem/validate_py.py .

Expected: 74 checks | 74 passed | 0 issues.

4. First scan

python3 .mind-mem/src/mind_mem/intel_scan.py .

Expected: 0 critical | 0 warnings on a fresh workspace.

5. Verify recall + capture

python3 .mind-mem/src/mind_mem/recall.py --query "test" --workspace .
# → No results found. (empty workspace — correct)

python3 .mind-mem/src/mind_mem/capture.py .
# → capture: no daily log for YYYY-MM-DD, nothing to scan (correct)

6. Add hooks (optional)

Option A: Claude Code hooks (recommended)

Merge into your .claude/hooks.json:

{
  "hooks": [
    {
      "event": "SessionStart",
      "command": "bash .mind-mem/hooks/session-start.sh"
    },
    {
      "event": "Stop",
      "command": "bash .mind-mem/hooks/session-end.sh"
    }
  ]
}

Option B: OpenClaw hooks (for OpenClaw 2026.2+)

cp -r .mind-mem/hooks/openclaw/mind-mem ~/.openclaw/hooks/mind-mem
openclaw hooks enable mind-mem

7. Smoke Test (optional)

bash .mind-mem/src/mind_mem/smoke_test.sh

Creates a temp workspace, runs init → validate → scan → recall → capture → pytest, then cleans up.


Health Summary

After setup, this is what a healthy workspace looks like:

$ python3 -m mind_mem.intel_scan .

mind-mem Intelligence Scan Report v2.0
Mode: detect_only

=== 1. CONTRADICTION DETECTION ===
  OK: No contradictions found among 25 signatures.

=== 2. DRIFT ANALYSIS ===
  OK: All active decisions referenced or exempt.
  INFO: Metrics: active_decisions=17, active_tasks=7, blocked=0,
        dead_decisions=0, incidents=3, decision_coverage=100%

=== 3. DECISION IMPACT GRAPH ===
  OK: Built impact graph: 11 decision(s) with edges.

=== 4. STATE SNAPSHOT ===
  OK: Snapshot saved.

=== 5. WEEKLY BRIEFING ===
  OK: Briefing generated.

TOTAL: 0 critical | 0 warnings | 16 info

Commands

Command

What it does

/scan

Run integrity scan — contradictions, drift, dead decisions, impact graph, snapshot, briefing

/apply

Review and apply proposals from scan results (dry-run first, then apply)

/recall <query>

Search across all memory files with ranked results (add --graph for cross-reference boosting)


Architecture

your-workspace/
├── mcp_server.py            # MCP server (FastMCP, 19 tools, 8 resources)
├── mind-mem.json             # Config
├── MEMORY.md                # Protocol rules
│
├── mind/                    # 17 MIND source files (.mind)
│   ├── bm25.mind           # BM25F scoring kernel
│   ├── rrf.mind            # Reciprocal Rank Fusion kernel
│   ├── reranker.mind        # Deterministic reranking
│   ├── abstention.mind      # Confidence gating
│   ├── ranking.mind         # Evidence ranking
│   ├── importance.mind      # A-MEM importance scoring
│   ├── category.mind        # Category relevance scoring
│   ├── recall.mind          # Combined recall scoring
│   ├── hybrid.mind          # BM25 + vector hybrid fusion
│   ├── rm3.mind             # RM3 pseudo-relevance feedback
│   ├── rerank.mind          # Score combination pipeline
│   ├── adversarial.mind     # Adversarial query detection
│   ├── temporal.mind        # Time-aware scoring
│   ├── prefetch.mind        # Context pre-assembly
│   ├── intent.mind          # Intent classification
│   └── cross_encoder.mind   # Cross-encoder blending
│
├── lib/                     # Compiled MIND kernels (optional)
│   └── libmindmem.so       # mindc output — not required for operation
│
├── decisions/
│   └── DECISIONS.md         # Formal decisions [D-YYYYMMDD-###]
├── tasks/
│   └── TASKS.md             # Tasks [T-YYYYMMDD-###]
├── entities/
│   ├── projects.md          # [PRJ-###]
│   ├── people.md            # [PER-###]
│   ├── tools.md             # [TOOL-###]
│   └── incidents.md         # [INC-###]
│
├── memory/
│   ├── YYYY-MM-DD.md        # Daily logs (append-only)
│   ├── intel-state.json     # Scanner state + metrics
│   └── maint-state.json     # Maintenance state
│
├── summaries/
│   ├── weekly/              # Weekly summaries
│   └── daily/               # Daily summaries
│
├── intelligence/
│   ├── CONTRADICTIONS.md    # Detected contradictions
│   ├── DRIFT.md             # Drift detections
│   ├── SIGNALS.md           # Auto-captured signals
│   ├── IMPACT.md            # Decision impact graph
│   ├── BRIEFINGS.md         # Weekly briefings
│   ├── AUDIT.md             # Applied proposal audit trail
│   ├── SCAN_LOG.md          # Scan history
│   ├── proposed/            # Staged proposals + resolution proposals
│   │   ├── DECISIONS_PROPOSED.md
│   │   ├── TASKS_PROPOSED.md
│   │   ├── EDITS_PROPOSED.md
│   │   └── RESOLUTIONS_PROPOSED.md
│   ├── applied/             # Snapshot archives (rollback)
│   └── state/snapshots/     # State snapshots
│
├── shared/                  # Multi-agent shared namespace
│   ├── decisions/
│   ├── tasks/
│   ├── entities/
│   └── intelligence/
│       └── LEDGER.md        # Cross-agent fact ledger
│
├── agents/                  # Per-agent private namespaces
│   └── <agent-id>/
│       ├── decisions/
│       ├── tasks/
│       └── memory/
│
├── mind-mem-acl.json        # Multi-agent access control
├── .mind-mem-wal/           # Write-ahead log (crash recovery)
│
└── src/mind_mem/
    ├── mind_ffi.py          # MIND FFI bridge (ctypes)
    ├── hybrid_recall.py     # Hybrid BM25+Vector+RRF orchestrator
    ├── block_metadata.py    # A-MEM metadata evolution
    ├── cross_encoder_reranker.py  # Optional cross-encoder
    ├── intent_router.py     # 9-type intent classification (adaptive)
    ├── recall.py            # BM25F + RM3 + graph scoring engine
    ├── recall_vector.py     # Vector/embedding backends
    ├── sqlite_index.py      # FTS5 + vector + metadata index
    ├── connection_manager.py # SQLite connection pool (WAL read/write separation)
    ├── block_store.py       # BlockStore protocol + MarkdownBlockStore
    ├── corpus_registry.py   # Central corpus path registry
    ├── abstention_classifier.py  # Adversarial abstention
    ├── evidence_packer.py   # Evidence assembly and ranking
    ├── intel_scan.py        # Integrity scanner
    ├── apply_engine.py      # Proposal apply engine (delta-based snapshots)
    ├── block_parser.py      # Markdown block parser (typed)
    ├── capture.py           # Auto-capture (26 patterns)
    ├── compaction.py        # Compaction/GC/archival
    ├── mind_filelock.py     # Cross-platform advisory file locking
    ├── observability.py     # Structured JSON logging + metrics
    ├── namespaces.py        # Multi-agent namespace & ACL
    ├── conflict_resolver.py # Automated conflict resolution
    ├── backup_restore.py    # WAL + backup/restore + JSONL export
    ├── transcript_capture.py  # Transcript JSONL signal extraction
    ├── validate.sh          # Structural validator (74+ checks)
    └── validate_py.py       # Structural validator (Python, cross-platform)

How It Compares

Quick Comparison

Feature

mind-mem

Mem0

Letta

Zep/Graphiti

Local-only

Yes

No (cloud API)

No (runtime)

No (Neo4j)

Zero infrastructure

Yes

No

No

No

Hybrid retrieval

BM25F + vector + RRF

Vector only

Hybrid

Graph + vector

Governance (propose/review/apply)

Yes

No

No

No

Contradiction detection

Yes

No

No

No

Tests

1,844

-

-

-

LoCoMo benchmark

67.3%

68.5%

74.0%

-

MCP tools

19

-

-

-

Core dependencies

0

Many

Many

Many

At a Glance

Tool

Strength

Trade-off

Mem0

Fast managed service, graph memory, multi-user scoping

Cloud-dependent, no integrity checking

Supermemory

Fastest retrieval (ms), auto-ingestion from Drive/Notion

Cloud-dependent, auto-writes without review

claude-mem

Purpose-built for Claude Code, ChromaDB vectors

Requires ChromaDB + Express worker, no integrity

Letta

Self-editing memory blocks, sleep-time compute, 74% LoCoMo

Full agent runtime (heavy), not just memory

Zep

Temporal knowledge graph, bi-temporal model, sub-second at scale

Cloud service, complex architecture

LangMem

Native LangChain/LangGraph integration

Tied to LangChain ecosystem

Cognee

Advanced chunking, web content bridging

Research-oriented, complex setup

Graphlit

Multimodal ingestion, semantic search, managed platform

Cloud-only, managed service

ClawMem

Full ML pipeline (cross-encoder + QMD + beam search)

4.5GB VRAM, 3 GPU processes required

MemU

Hierarchical 3-layer memory, multimodal ingestion, LLM-based retrieval

Requires LLM for extraction and retrieval, no hybrid search

mind-mem

Integrity + governance + zero core deps + hybrid search + MIND kernels + 19 MCP tools

Lexical recall by default (vector/CE optional)

Full Feature Matrix

Compared against every major memory solution for AI agents (as of 2026):

Mem0

Supermemory

claude-mem

Letta

Zep

LangMem

Cognee

Graphlit

ClawMem

MemU

mind-mem

Recall

Vector

Cloud

Cloud

Chroma

Yes

Yes

Yes

Yes

Yes

Yes

Optional

Lexical

Filter

BM25

BM25F

Graph

Yes

Yes

Yes

Yes

Beam

2-hop

Hybrid + RRF

Part

Yes

Yes

Yes

Yes

Yes

Cross-encoder

qwen3 0.6B

MiniLM 80MB

Intent routing

Yes

9 types

Query expansion

QMD 1.7B

RM3 (zero-dep)

Persistence

Structured

JSON

JSON

SQL

Blk

Grph

KV

Grph

Grph

SQL

Markdown

Markdown

Entities

Yes

Yes

Yes

Yes

Yes

Yes

Yes

Yes

Yes

Temporal

Yes

Yes

Supersede

Yes

Yes

Yes

Append-only

Yes

A-MEM metadata

Yes

Yes

Integrity

Contradictions

Yes

Drift detection

Yes

Validation

74+ rules

Impact graph

Yes

Coverage

Yes

Multi-agent

Yes

ACL-based

Conflict res.

Automatic

WAL/crash

Yes

Backup/restore

Yes

Abstention

Yes

Governance

Auto-capture

Auto

Auto

Auto

Self

Ext

Ext

Ext

Ing

Auto

LLM Ext

Propose

Proposal queue

Yes

Rollback

Yes

Mode governance

3 modes

Audit trail

Part

Full

Operations

Local-only

Yes

Yes

Yes

Yes

Zero core deps

Yes

No daemon

Yes

Yes

Yes

GPU required

4.5GB

No

No

Git-friendly

Part

Yes

Yes

MCP server

19 tools

MIND kernels

16 source

The Gap mind-mem Fills

Every tool above does storage + retrieval. None of them answer:

  • "Do any of my decisions contradict each other?"

  • "Which decisions are active but nobody references anymore?"

  • "Did I make a decision in chat that was never formalized?"

  • "What's the downstream impact if I change this decision?"

  • "Is my memory state structurally valid right now?"

mind-mem focuses on memory governance and integrity — the critical layer most memory systems ignore entirely.

Why Plain Files Outperform Fancy Retrieval

Letta's August 2025 analysis showed that a plain-file baseline (full conversations stored as files + agent filesystem tools) scored 74.0% on LoCoMo with gpt-4o-mini — beating Mem0's top graph variant at 68.5%. Key reasons:

  • LLMs excel at tool-based retrieval. Agents can iteratively query/refine file searches better than single-shot vector retrieval that might miss subtle connections.

  • Benchmarks reward recall + reasoning over storage sophistication. Strong judge LLMs handle the rest once relevant chunks are loaded.

  • Overhead hurts. Specialized pipelines introduce failure modes (bad embeddings, chunking errors, stale indexes) that simple file access avoids.

  • For text-heavy agentic use cases, "how well the agent manages context" > "how smart the retrieval index is."

mind-mem's deterministic retrieval pipeline validates these findings: 67.3% on LoCoMo with zero dependencies, no embeddings, and no vector database — within 1.2pp of Mem0's graph-based approach. The key insight: treating retrieval as a reasoning pipeline (wide candidate pool → deterministic rerank → context packing) closes most of the gap without any ML infrastructure. Unlike plain-file baselines, mind-mem adds integrity checking, governance, and agent-agnostic shared memory via MCP that no other system provides.


Recall

Default: BM25 Hybrid

python3 -m mind_mem.recall --query "authentication" --workspace .
python3 -m mind_mem.recall --query "auth" --json --limit 5 --workspace .
python3 -m mind_mem.recall --query "deadline" --active-only --workspace .

BM25F scoring (k1=1.2, b=0.75) with per-field weighting, bigram phrase matching, overlapping sentence chunking, and query-type-aware parameter tuning. Searches across all structured files.

BM25F field weighting: Terms in Statement fields score 3x higher than terms in Context (0.5x). This naturally prioritizes core content over auxiliary metadata.

RM3 query expansion: Pseudo-relevance feedback from top-k initial results. JM-smoothed language model extracts expansion terms, interpolated with the original query at configurable alpha. Falls back to static synonyms for adversarial queries.

Adversarial abstention: Deterministic pre-LLM confidence gate. Computes confidence from entity overlap, BM25 score, speaker coverage, evidence density, and negation asymmetry. Below threshold → forces abstention.

Stemming: "queries" matches "query", "deployed" matches "deployment". Simplified Porter stemmer with zero dependencies.

Hybrid Search (BM25 + Vector + RRF)

{
  "recall": {
    "backend": "hybrid",
    "vector_enabled": true,
    "rrf_k": 60,
    "bm25_weight": 1.0,
    "vector_weight": 1.0
  }
}

Thread-parallel BM25 and vector retrieval fused via RRF: score(doc) = bm25_w / (k + bm25_rank) + vec_w / (k + vec_rank). Deduplicates by block ID. Falls back to BM25-only when vector backend is unavailable.

Graph-Based (2-hop cross-reference boost)

python3 -m mind_mem.recall --query "database" --graph --workspace .

2-hop graph traversal: 1-hop neighbors get 0.3x score boost, 2-hop get 0.1x (tagged [graph]). Surfaces structurally connected blocks via AlignsWith, Dependencies, Supersedes, Sources, and ConstraintSignature scopes. Auto-enabled for multi-hop queries.

Vector (pluggable)

{
  "recall": {
    "backend": "vector",
    "vector_enabled": true,
    "vector_model": "all-MiniLM-L6-v2",
    "onnx_backend": true
  }
}

Supports ONNX inference (local, no server) or cloud embeddings. Falls back to BM25 automatically if unavailable.


MIND Kernels

mind-mem includes 17 .mind kernel source files — numerical hot paths written in the MIND programming language. The MIND kernel is optional. mind-mem works identically without it (pure Python fallback). With it, scoring runs at native speed with compile-time tensor shape verification.

Compilation

Requires the MIND compiler (mindc). See mindlang.dev for installation.

# Compile all kernels to a single shared library
mindc mind/*.mind --emit=shared -o lib/libmindmem.so

# Or compile individually for testing
mindc mind/bm25.mind --emit=shared -o lib/libbm25.so

Kernel Index

File

Functions

Purpose

bm25.mind

bm25f_doc, bm25f_batch, apply_recency, apply_graph_boost

BM25F scoring with field boosts

rrf.mind

rrf_fuse, rrf_fuse_three

Reciprocal Rank Fusion

reranker.mind

date_proximity_score, category_boost, negation_penalty, rerank_deterministic

Deterministic reranking

rerank.mind

rerank_scores

Score combination pipeline

abstention.mind

entity_overlap, confidence_score

Confidence gating

ranking.mind

weighted_rank, top_k_mask

Evidence ranking

importance.mind

importance_score

A-MEM importance scoring

category.mind

category_affinity, query_category_relevance, category_assign

Category distillation scoring

prefetch.mind

prefetch_score, prefetch_select

Signal-based context pre-assembly

recall.mind

recall_score

Combined recall scoring

hybrid.mind

hybrid_fuse

BM25 + vector hybrid fusion

rm3.mind

rm3_weight

RM3 pseudo-relevance feedback

adversarial.mind

adversarial_gate

Adversarial query detection

temporal.mind

temporal_decay

Time-aware scoring

intent.mind

intent_params

Intent classification parameters

cross_encoder.mind

ce_blend

Cross-encoder blending configuration

Performance

 

Function

N=100

N=1,000

N=5,000

rrf_fuse

10.8x

69.0x

72.5x

bm25f_batch

13.2x

113.8x

193.1x

negation_penalty

3.3x

7.0x

18.4x

date_proximity

10.7x

15.3x

26.9x

category_boost

3.3x

19.8x

17.7x

importance_batch

22.3x

46.2x

48.6x

confidence_score

0.9x

0.8x

0.9x

top_k_mask

3.1x

8.1x

11.8x

weighted_rank

5.1x

26.6x

121.8x

Overall

49.0x

49x faster end-to-end at production scale (N=5,000). Individual kernels reach up to 193x speedup. The compiled library includes 14 runtime protection layers with near-zero overhead.

FFI Bridge

The compiled .so exposes a C99-compatible ABI. Python calls via ctypes through src/mind_mem/mind_ffi.py:

from mind_ffi import get_kernel, is_available, is_protected

if is_available():
    kernel = get_kernel()
    scores = kernel.rrf_fuse_py(bm25_ranks, vec_ranks, k=60.0)
    print(f"Protected: {is_protected()}")  # True with FORTRESS build

Without MIND

If lib/libmindmem.so is not present, mind-mem uses pure Python implementations. The Python fallback produces identical results (within f32 epsilon). No functionality is lost — MIND is a performance optimization, not a requirement.


Auto-Capture

Session end
    ↓
capture.py scans daily log (or --scan-all for batch)
    ↓
Detects decision/task language (26 patterns, 3 confidence levels)
    ↓
Extracts structured metadata (subject, object, tags)
    ↓
Classifies confidence (high/medium/low → P1/P2/P3)
    ↓
Writes to intelligence/SIGNALS.md ONLY
    ↓
User reviews signals
    ↓
/apply promotes to DECISIONS.md or TASKS.md

Batch scanning: python3 -m mind_mem.capture . --scan-all scans the last 7 days of daily logs.

Safety guarantee: capture.py never writes to decisions/ or tasks/ directly. All signals must go through the apply engine.


Multi-Agent Memory

Namespace Setup

python3 -m mind_mem.namespaces workspace/ --init coder-1 reviewer-1

Creates shared/ (visible to all) and agents/coder-1/, agents/reviewer-1/ (private) directories with ACL config.

Access Control

{
  "default_policy": "read",
  "agents": {
    "coder-1": {"namespaces": ["shared", "agents/coder-1"], "write": ["agents/coder-1"], "read": ["shared"]},
    "reviewer-*": {"namespaces": ["shared"], "write": [], "read": ["shared"]},
    "*": {"namespaces": ["shared"], "write": [], "read": ["shared"]}
  }
}

Shared Fact Ledger

High-confidence facts proposed to shared/intelligence/LEDGER.md become visible to all agents after review. Append-only with dedup and file locking.

Conflict Resolution

python3 -m mind_mem.conflict_resolver workspace/ --analyze
python3 -m mind_mem.conflict_resolver workspace/ --propose

Graduated resolution: confidence priority > scope specificity > timestamp priority > manual fallback.

Transcript Capture

python3 -m mind_mem.transcript_capture workspace/ --transcript path/to/session.jsonl
python3 -m mind_mem.transcript_capture workspace/ --scan-recent --days 3

Scans Claude Code JSONL transcripts for user corrections, convention discoveries, and architectural decisions. 16 patterns with confidence classification.

Backup & Restore

python3 -m mind_mem.backup_restore backup workspace/ --output backup.tar.gz
python3 -m mind_mem.backup_restore export workspace/ --output export.jsonl
python3 -m mind_mem.backup_restore restore workspace/ --input backup.tar.gz
python3 -m mind_mem.backup_restore wal-replay workspace/

Governance Modes

Mode

What it does

When to use

detect_only

Scan + validate + report only

Start here. First week after install.

propose

Report + generate fix proposals in proposed/

After a clean observation week with zero critical issues.

enforce

Bounded auto-supersede + self-healing within constraints

Production mode. Requires explicit opt-in.

Recommended rollout:

  1. Install → run in detect_only for 7 days

  2. Review scan logs → if clean, switch to propose

  3. Triage proposals for 2-3 weeks → if confident, enable enforce


Block Format

All structured data uses a simple, parseable markdown format:

[D-20260213-001]
Date: 2026-02-13
Status: active
Statement: Use PostgreSQL for the user database
Tags: database, infrastructure
Rationale: Better JSON support than MySQL for our use case
ConstraintSignatures:
- id: CS-db-engine
  domain: infrastructure
  subject: database
  predicate: engine
  object: postgresql
  modality: must
  priority: 9
  scope: {projects: [PRJ-myapp]}
  evidence: Benchmarked JSON performance
  axis:
    key: database.engine
  relation: standalone
  enforcement: structural

Blocks are parsed by block_parser.py — a zero-dependency markdown parser that extracts [ID] headers and Key: Value fields into structured dicts.


Configuration

All settings in mind-mem.json (created by init_workspace.py):

{
  "version": "1.9.1",
  "workspace_path": ".",
  "auto_capture": true,
  "auto_recall": true,
  "governance_mode": "detect_only",
  "recall": {
    "backend": "bm25",
    "rrf_k": 60,
    "bm25_weight": 1.0,
    "vector_weight": 1.0,
    "vector_model": "all-MiniLM-L6-v2",
    "vector_enabled": false,
    "onnx_backend": false
  },
  "proposal_budget": {
    "per_run": 3,
    "per_day": 6,
    "backlog_limit": 30
  },
  "compaction": {
    "archive_days": 90,
    "snapshot_days": 30,
    "log_days": 180,
    "signal_days": 60
  },
  "scan_schedule": "daily"
}

Key

Default

Description

version

"1.9.1"

Config file version

auto_capture

true

Run capture engine on session end

auto_recall

true

Show recall context on session start

governance_mode

"detect_only"

Governance mode (detect_only, propose, enforce)

recall.backend

"scan"

"scan" (BM25), "hybrid" (BM25+Vector+RRF), or "vector"

recall.rrf_k

60

RRF fusion parameter k

recall.bm25_weight

1.0

BM25 weight in RRF fusion

recall.vector_weight

1.0

Vector weight in RRF fusion

recall.vector_model

"all-MiniLM-L6-v2"

Embedding model for vector search

recall.vector_enabled

false

Enable vector search backend

recall.onnx_backend

false

Use ONNX for local embeddings (no server needed)

proposal_budget.per_run

3

Max proposals generated per scan

proposal_budget.per_day

6

Max proposals per day

proposal_budget.backlog_limit

30

Max pending proposals before pausing

compaction.archive_days

90

Archive completed blocks older than N days

compaction.snapshot_days

30

Remove apply snapshots older than N days

compaction.log_days

180

Archive daily logs older than N days

compaction.signal_days

60

Remove resolved/rejected signals older than N days

scan_schedule

"daily"

"daily" or "manual"


MCP Server

mind-mem ships with a Model Context Protocol server that exposes memory as resources and tools to any MCP-compatible client.

Install

pip install fastmcp
./install.sh --all

Configures all detected clients automatically. See Quick Start.

Manual Setup

For Claude Code, Claude Desktop, Cursor, Windsurf, and Gemini CLI, add to the respective JSON config under mcpServers:

{
  "mcpServers": {
    "mind-mem": {
      "command": "python3",
      "args": ["/path/to/mind-mem/mcp_server.py"],
      "env": {"MIND_MEM_WORKSPACE": "/path/to/your/workspace"}
    }
  }
}

Client

Config File

Claude Code CLI

~/.claude/mcp.json

Claude Desktop

~/.config/Claude/claude_desktop_config.json

Gemini CLI

~/.gemini/settings.json

Cursor

~/.cursor/mcp.json

Windsurf

~/.codeium/windsurf/mcp_config.json

For Codex CLI (TOML format), add to ~/.codex/config.toml:

[mcp_servers.mind-mem]
command = "python3"
args = ["/path/to/mind-mem/mcp_server.py"]

[mcp_servers.mind-mem.env]
MIND_MEM_WORKSPACE = "/path/to/your/workspace"

For Zed, add to ~/.config/zed/settings.json under context_servers:

{
  "context_servers": {
    "mind-mem": {
      "command": {
        "path": "python3",
        "args": ["/path/to/mind-mem/mcp_server.py"],
        "env": {"MIND_MEM_WORKSPACE": "/path/to/your/workspace"}
      }
    }
  }
}

Direct (stdio / HTTP)

# stdio transport (default)
MIND_MEM_WORKSPACE=/path/to/workspace python3 mcp_server.py

# HTTP transport (multi-client / remote)
MIND_MEM_WORKSPACE=/path/to/workspace python3 mcp_server.py --transport http --port 8765

Resources (read-only)

URI

Description

mind-mem://decisions

Active decisions

mind-mem://tasks

All tasks

mind-mem://entities/{type}

Entities (projects, people, tools, incidents)

mind-mem://signals

Auto-captured signals pending review

mind-mem://contradictions

Detected contradictions

mind-mem://health

Workspace health summary

mind-mem://recall/{query}

BM25 recall search results

mind-mem://ledger

Shared fact ledger (multi-agent)

Tools (21)

Tool

Description

recall

Search memory with BM25 (query, limit, active_only)

propose_update

Propose a decision/task — writes to SIGNALS.md only

approve_apply

Apply a staged proposal (dry_run=True by default)

rollback_proposal

Rollback an applied proposal by receipt timestamp

scan

Run integrity scan (contradictions, drift, signals)

list_contradictions

List contradictions with auto-resolution analysis

hybrid_search

Hybrid BM25+Vector search with RRF fusion

find_similar

Find blocks similar to a given block

intent_classify

Classify query intent (9 types with parameter recommendations)

index_stats

Index statistics, MIND kernel availability, block counts

retrieval_diagnostics

Pipeline rejection rates, intent histogram, hard negatives

reindex

Rebuild FTS5 index (optionally including vectors)

memory_evolution

View/trigger A-MEM metadata evolution for a block

list_mind_kernels

List available MIND kernel configurations

get_mind_kernel

Read a specific MIND kernel configuration as JSON

category_summary

Category summaries relevant to a given topic

prefetch

Pre-assemble context from recent conversation signals

delete_memory_item

Delete a memory block by ID (admin-scope)

export_memory

Export workspace as JSONL (user-scope)

calibration_feedback

Submit quality feedback for a retrieved block (thumbs up/down)

calibration_stats

View per-block and global calibration statistics

Token Auth (HTTP)

MIND_MEM_TOKEN=your-secret python3 mcp_server.py --transport http --port 8765

Safety Guarantees

  • propose_update never writes to DECISIONS.md or TASKS.md. All proposals go to SIGNALS.md.

  • approve_apply defaults to dry_run=True. Creates a snapshot before applying for rollback.

  • All resources are read-only. No MCP client can mutate source of truth through resources.

  • Namespace-aware. Multi-agent workspaces scope resources by agent ACL.


Security

Threat Model

What we protect

How

Memory integrity

74+ structural checks, ConstraintSignature validation

Accidental overwrites

Proposal-based mutations only (never direct writes)

Rollback safety

Snapshot before every apply, atomic os.replace()

Symlink attacks

Symlink detection in restore paths

Path traversal

All paths resolved via os.path.realpath(), workspace-relative only

What we do NOT protect against

Why

Malicious local user

Single-user CLI tool — filesystem access = data access

Network attacks

No network calls, no listening ports, no telemetry

Encrypted storage

Files are plaintext Markdown — use disk encryption if needed

No Network Calls

mind-mem makes zero network calls from its core. No telemetry, no phoning home, no cloud dependencies. Optional features (vector embeddings, cross-encoder) may download models on first use.


Requirements

  • Python 3.10+

  • No external packages — stdlib only for core functionality

Optional Dependencies

Package

Purpose

Install

fastmcp

MCP server

pip install mind-mem[mcp]

onnxruntime + tokenizers

Local vector embeddings

pip install mind-mem[embeddings]

sentence-transformers

Cross-encoder reranking

pip install mind-mem[cross-encoder]

ollama

LLM extraction (local)

pip install ollama

Platform Support

Platform

Status

Notes

Linux

Full

Primary target

macOS

Full

POSIX-compliant shell scripts

Windows (WSL/Git Bash)

Full

Use WSL2 or Git Bash for shell hooks

Windows (native)

Python only

Use validate_py.py; hooks require WSL


Troubleshooting

Problem

Solution

validate.sh says "No mind-mem.json found"

Run in a workspace, not the repo root. Run init_workspace.py first.

recall returns no results

Workspace is empty. Add decisions/tasks first.

capture says "no daily log"

No memory/YYYY-MM-DD.md for today. Write something first.

intel_scan finds 0 contradictions

Good — no conflicting decisions.

Tests fail on Windows

Use validate_py.py instead of validate.sh. Hooks require WSL.

MIND kernel not loading

Compile with mindc mind/*.mind --emit=shared -o lib/libmindmem.so. Or ignore — pure Python works identically.

FAQ

No results from recall? Check that the workspace path is correct and points to an initialized workspace containing decisions, tasks, or entities. If the FTS5 index is stale or missing, run the reindex MCP tool to rebuild it.

MCP connection failed? Verify that fastmcp is installed (pip install fastmcp). Check the transport configuration in your client's MCP config (stdio vs HTTP). Ensure the MIND_MEM_WORKSPACE environment variable points to a valid workspace directory.

MIND kernels not loading? Run bash src/mind_mem/build.sh to compile the MIND source files (requires mindc). If the MIND compiler is not available, mind-mem automatically uses the pure Python fallback with identical results.

Index corrupt? Run the reindex MCP tool, or from the command line: python3 -m mind_mem.sqlite_index --rebuild --workspace /path/to/workspace. This drops and recreates the FTS5 index from all workspace files.


Specification

For the formal grammar, invariant rules, state machine, and atomicity guarantees, see SPEC.md.


Contributing

Contributions welcome. Please open an issue first to discuss what you'd like to change.

See CONTRIBUTING.md for guidelines.


License

Apache 2.0 — Copyright 2026 STARGA Inc and contributors.

-
security - not tested
A
license - permissive license
-
quality - not tested

Resources

Unclaimed servers have limited discoverability.

Looking for Admin?

If you are the server author, to access and configure the admin panel.

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/star-ga/mind-mem'

If you have feedback or need assistance with the MCP directory API, please join our Discord server