Project Tessera is a local, private AI knowledge layer that captures, stores, and serves knowledge across AI tools and sessions. All processing (embeddings, storage) runs locally via LanceDB and fastembed/ONNX — no external API calls.
Search & Retrieval
search_documents— Hybrid semantic + keyword search across indexed documents (PRDs, decision logs, session logs, etc.), with optional filtering by project or typelist_sources— List all indexed files in the workspaceread_file— Read the full contents of any file
Memory & Learning
remember— Save knowledge, decisions, or preferences for cross-session persistencerecall— Search and retrieve memories from past conversationslearn— Auto-save and immediately index new insights for future search
File Organization
organize_files— Move, rename, or archive workspace filessuggest_cleanup— Detect and suggest cleanup for clutter, backups, large files, and empty directories
Project & Decision Tracking
project_status— Get recent changes and file statisticsextract_decisions— Pull past decisions from session and decision logsaudit_prd— Audit a PRD for quality and completeness across a 13-section structure
Knowledge Graphs
knowledge_graph— Generate a Mermaid diagram of relationships between documents and conceptsexplore_connections— Show connections for a specific document or concept
Indexing & Sync
ingest_documents— Full rebuild of the local vector indexsync_documents— Incremental sync of only new, changed, or deleted files
Integrates with Claude Desktop via MCP, with planned HTTP API support for other platforms (ChatGPT, Gemini).
Tessera
Every AI conversation produces knowledge. When the session ends, it's gone. Tessera keeps it.
One knowledge base shared across Claude, ChatGPT, Gemini, and Copilot. Runs locally. No API keys, no Docker, no data leaving your machine.
pip install project-tessera
tessera setup
# Done. Claude Desktop now has persistent memory + document search.Why Tessera over alternatives
Tessera | Mem0 | Basic Memory | mcp-memory-service | |
Works without API keys | Yes | No (needs OpenAI) | Yes | Partial |
Works without Docker | Yes | No | Yes | No |
Document search (40+ types) | Yes | No | Markdown only | No |
ChatGPT live integration (Actions) | Yes | No | No | No |
Contradiction detection | Yes | No | No | No |
Memory confidence scoring | Yes | No | No | No |
Encrypted vault (AES-256) | Yes | No | No | No |
HTTP API for non-MCP tools | 54 endpoints | Yes | No | Yes |
Auto-learning from conversations | Yes | Yes | No | No |
MCP tools | 58 | ~10 | ~15 | 24 |
The short version
Most memory tools store text and search it. Tessera does that, plus:
Cross-AI: ChatGPT calls Tessera's API through Custom GPT Actions. One knowledge base, two AIs reading and writing to it live.
Self-maintaining: finds contradictions between old and new memories, scores confidence by reinforcement frequency, flags stale knowledge, auto-merges near-duplicates.
Zero infrastructure:
pip installand go. LanceDB and fastembed are embedded -- no Docker, no database server, no API keys.Encrypted: set
TESSERA_VAULT_KEYand all memories are AES-256-CBC encrypted at rest.
Related MCP server: MCP VectorStore Server
Architecture
How search works (query path)
User asks: "What did we decide about the database?"
|
v
+-----------------------+
| Query Processing |
| Multi-angle decomp | "database decision"
| (2-4 perspectives) | "database", "decision"
+-----------------------+ "decision about database"
|
+-------------+-------------+
| |
v v
+------------------+ +------------------+
| Vector Search | | Keyword Search |
| (LanceDB) | | (FTS index) |
| 384-dim MiniLM | | BM25 scoring |
+------------------+ +------------------+
| |
+-------------+-------------+
|
v
+-----------------------+
| Reranking |
| 70% semantic weight | LinearCombinationReranker
| 30% keyword weight | + version-aware scoring
+-----------------------+
|
v
+-----------------------+
| Result Assembly |
| Dedup (content hash) | 2-pass deduplication
| Verdict labels | found / weak / none
| Cache (60s TTL) |
+-----------------------+
|
v
Top-K results with
confidence scoresHow ingestion works (ingest path)
Documents: .md .pdf .docx .xlsx .py .ts .go ... (40+ types)
|
v
+-----------------------+
| File Type Router |
| Markdown, CSV, XLSX | Type-specific parsers
| Code, PDF, Images | with metadata extraction
+-----------------------+
|
v
+-----------------------+
| Chunking Engine |
| 1024 tokens/chunk | Sentence-boundary aware
| 100 token overlap | Heading-preserving
+-----------------------+
|
v
+-----------------------+
| Local Embedding |
| fastembed/ONNX | paraphrase-multilingual
| 384 dimensions | MiniLM-L12-v2
| No API calls | 101 languages
+-----------------------+
|
+-------------+-------------+
| |
v v
+------------------+ +------------------+
| LanceDB | | SQLite |
| Vector storage | | File metadata |
| Columnar format | | Search analytics|
| Zero-config | | Interaction log |
+------------------+ +------------------+System overview
+--------------------------------------------+
| src/core.py |
| 58 orchestration functions |
| 69 specialized modules, 31k LOC |
+--------------------------------------------+
/ | \
+---------------+ +-------------------+ +--------------+
| MCP Server | | HTTP API Server | | CLI |
| Claude Desktop| | FastAPI + Swagger | | 11 commands |
| 58 tools | | 54 endpoints | | setup, sync |
| stdio | | port 8394 | | ingest, api |
+---------------+ +-------------------+ +--------------+
| | |
v v v
+------------------------------------------------------------+
| Storage Layer |
| LanceDB SQLite Filesystem |
| (vectors) (metadata, (memories as .md, |
| analytics, encrypted with |
| interactions) AES-256-CBC) |
| |
| fastembed/ONNX: local embedding, no API keys |
| 101 languages, 384-dim vectors, ~220MB model |
+------------------------------------------------------------+Get started
1. Install
pip install project-tesseraOr with uv:
uvx --from project-tessera tessera setup2. Setup
tessera setupCreates workspace config, downloads embedding model (~220MB, first time only), configures Claude Desktop.
3. Restart Claude Desktop
Ask Claude about your documents. It searches automatically.
Use with ChatGPT (Custom GPT Actions)
tessera api # Start REST API on localhost:8394
ngrok http 8394 # Expose to the internet
# Then create a Custom GPT with the Actions spec from /chatgpt-actions/openapi.jsonFull setup guide at http://127.0.0.1:8394/chatgpt-actions/setup. Swagger docs at http://127.0.0.1:8394/docs.
How it works
Hybrid search with reranking
Every search goes through four stages:
Query decomposition -- the query is split into 2-4 search angles (core keywords, individual terms, reversed emphasis)
Hybrid retrieval -- vector similarity (LanceDB) and keyword matching (FTS/BM25) run in parallel
Reranking -- a LinearCombinationReranker merges the two result sets (70% semantic, 30% keyword weight)
Verdict scoring -- each result gets a label:
confident match(>= 45%),possible match(25-45%), orlow relevance(< 25%)
When multiple versions of the same document exist, Tessera prefers the latest.
Cross-session memory
# Via MCP (Claude)
"Remember that we chose PostgreSQL for the production database"
# Via HTTP API (ChatGPT, Gemini, scripts)
curl -X POST http://127.0.0.1:8394/remember \
-H "Content-Type: application/json" \
-d '{"content": "Use PostgreSQL for production", "tags": ["db", "architecture"]}'Each memory gets a category (decision, preference, or fact), is checked for duplicates against existing memories (cosine similarity, 0.92 threshold), and receives a confidence score -- weighted by repetition (35%), recency (25%), source diversity (20%), and category (20%). Set TESSERA_VAULT_KEY to encrypt all memories with AES-256-CBC.
Auto-learning
Tessera picks up decisions, preferences, and facts from your conversations without being asked. toggle_auto_learn turns it on or off; review_learned shows what it caught.
Contradiction detection
Memories contradict each other over time. Tessera finds them:
CONTRADICTION (HIGH severity):
"We decided to use PostgreSQL" (2026-03-01)
vs
"Switched to MongoDB for the main database" (2026-03-10)
The newer memory (2026-03-10) likely reflects the current state.Works with both English and Korean negation patterns.
Cross-AI: ChatGPT Custom GPT Actions
ChatGPT talks to Tessera's HTTP API through Custom GPT Actions. No export/import step -- it reads and writes the same knowledge base Claude uses through MCP, in real time.
# 1. Start Tessera API + tunnel
tessera api
ngrok http 8394
# 2. Get the OpenAPI spec for your Custom GPT
curl https://your-tunnel.ngrok-free.app/chatgpt-actions/openapi.json?server_url=https://your-tunnel.ngrok-free.app
# 3. Get the GPT instruction template
curl https://your-tunnel.ngrok-free.app/chatgpt-actions/instructions
# 4. Full setup guide
curl https://your-tunnel.ngrok-free.app/chatgpt-actions/setupCreate a Custom GPT, paste the instructions, import the OpenAPI spec as an Action. ChatGPT can then search your documents, save memories, and recall past decisions from the same knowledge base.
You can also import past ChatGPT conversations to extract knowledge from them:
curl -X POST http://127.0.0.1:8394/import-conversations \
-H "Content-Type: application/json" \
-d '{"data": "<ChatGPT export JSON>", "source": "chatgpt"}'Export as Obsidian vault (wikilinks), Markdown, CSV, or JSON:
curl http://127.0.0.1:8394/export?format=obsidianMemory health
Each memory is healthy, stale (90+ days without reinforcement), or orphaned (no metadata, no category). The health report tells you what to clean up and tracks growth over time.
Plugin hooks
Run your own scripts when things happen:
# workspace.yaml
hooks:
on_memory_created:
- script: ./notify-slack.sh
on_contradiction_found:
- script: ./alert.py7 event types: on_memory_created, on_memory_deleted, on_search, on_session_start, on_session_end, on_ingest_complete, on_contradiction_found.
Supported file types (40+)
Category | Extensions | Install |
Documents |
| included |
Office |
|
|
Code |
| included |
Config |
| included |
Web |
| included |
Images |
|
|
MCP tools (58)
Tool | What it does |
| Semantic + keyword hybrid search across all docs |
| Search documents AND memories in one call |
| Full file view (CSV as table, XLSX per sheet) |
| Read any file's full content |
| See what's indexed |
Tool | What it does |
| Save knowledge that persists across sessions |
| Search past memories with date/category filters |
| Save and immediately index new knowledge |
| Browse saved memories |
| Delete a specific memory |
| Batch export all memories as JSON |
| Batch import memories from JSON |
| List all unique tags with counts |
| Filter memories by specific tag |
| List auto-detected categories (decision/preference/fact) |
| Filter memories by category |
| Find documents similar to a given file |
| Build a Mermaid diagram of document relationships |
Tool | What it does |
| Extract and save knowledge from the current session |
| Turn auto-learning on/off or check status |
| Review recently auto-learned memories |
| View tool calls from current/past sessions |
| Session history with interaction counts |
Tool | What it does |
| How your decisions changed over time, by topic |
| Pack the best context into a token budget |
| Query suggestions based on your past searches |
| Cluster memories by topic with Mermaid mindmap |
| Aggregate statistics (categories, tags, growth) |
| Auto-built profile (language, preferences, expertise) |
| Show connections around a specific topic |
Tool | What it does |
| Breaks a query into 2-4 angles, searches each, merges best results |
| Multi-angle memory recall with verdict labels |
| Find conflicting memories with severity rating |
| How reliable is each memory (repetition, recency, source diversity) |
| Which memories are healthy, stale, or orphaned |
| See what hooks are registered |
Tool | What it does |
| Export memories in ChatGPT/Gemini-readable format |
| Import memories from external AI tools |
| Extract knowledge from ChatGPT/Claude conversation exports |
| Export as Obsidian (wikilinks), Markdown, CSV, or JSON |
ChatGPT connects via Custom GPT Actions (HTTP API). See /chatgpt-actions/setup for the full guide.
Tool | What it does |
| Check AES-256 encryption status |
| Upgrade data from older schema versions |
Tool | What it does |
| Index documents (first-time or full rebuild) |
| Incremental sync (only changed files) |
| Recent changes per project |
| Find past decisions from logs |
| Check PRD quality (13-section structure) |
| Move, rename, archive files |
| Detect backup files, empty dirs, misplaced files |
| Server health: tracked files, sync history, cache |
| Full workspace diagnostics |
| Search usage patterns, top queries, response times |
| Detect stale documents older than N days |
HTTP API (54 endpoints)
pip install project-tessera[api]
tessera api # http://127.0.0.1:8394Swagger UI at http://127.0.0.1:8394/docs. Optional auth via TESSERA_API_KEY env var.
Method | Path | What it does |
GET |
| Health check |
GET |
| Version info |
POST |
| Semantic + keyword search |
POST |
| Search docs + memories |
POST |
| Save a memory |
POST |
| Search memories with filters |
POST |
| Save and index knowledge |
GET |
| List memories |
DELETE |
| Delete a memory |
GET |
| List categories |
GET |
| Filter by category |
GET |
| List tags |
GET |
| Filter by tag |
POST |
| Build token-budgeted context |
GET |
| Decision evolution |
GET |
| Query suggestions |
GET |
| Topic clusters |
GET |
| Stats dashboard |
POST |
| Multiple operations in one call |
GET |
| Export as Obsidian/MD/CSV/JSON |
GET |
| Export for ChatGPT/Gemini |
POST |
| Import from ChatGPT/Gemini |
POST |
| Import past conversations |
POST |
| Run data migration |
GET |
| Encryption status |
GET |
| User profile |
GET |
| Server status |
GET |
| Workspace diagnostics |
POST |
| Multi-angle document search |
POST |
| Multi-angle memory recall |
GET |
| Detect conflicting memories |
GET |
| Memory reliability scores |
GET |
| Memory health analytics |
GET |
| List plugin hooks |
GET |
| Search entity knowledge graph |
POST |
| Mermaid diagram from entities |
GET |
| Find similar memory clusters |
POST |
| Merge similar memories |
GET |
| Web dashboard (dark theme, entity graph, stats) |
POST |
| Auto-merge near-duplicate memories |
POST |
| Flag old or low-quality memories |
GET |
| Age distribution and at-risk counts |
GET |
| Setup code for LangChain, CrewAI, AutoGen |
POST |
| Classify, tag, deduplicate, and clean up memories |
GET |
| Trending topics, decision patterns, hidden connections |
GET |
| OpenAPI spec for Custom GPT Actions |
GET |
| GPT instruction template |
GET |
| ChatGPT integration setup guide |
Quick examples
# Search documents
curl -X POST http://127.0.0.1:8394/search \
-H "Content-Type: application/json" \
-d '{"query": "database architecture", "top_k": 5}'
# Save a memory
curl -X POST http://127.0.0.1:8394/remember \
-H "Content-Type: application/json" \
-d '{"content": "Use PostgreSQL for production", "tags": ["db"]}'
# Export for ChatGPT
curl http://127.0.0.1:8394/export-for-ai?target=chatgpt
# Batch (multiple operations, single request)
curl -X POST http://127.0.0.1:8394/batch \
-H "Content-Type: application/json" \
-d '{"operations": [{"method": "search", "params": {"query": "test"}}, {"method": "knowledge_stats"}]}'CLI (11 commands)
tessera setup # One-command setup (config + model download + Claude Desktop)
tessera init # Interactive setup
tessera ingest # Index all document sources
tessera sync # Re-index changed files only
tessera serve # Start MCP server (stdio)
tessera api # Start HTTP API server (port 8394)
tessera migrate # Upgrade data schema
tessera check # Workspace health diagnostics
tessera status # Project status summary
tessera install-mcp # Configure Claude Desktop
tessera version # Show versionClaude Desktop config
With uvx (recommended):
{
"mcpServers": {
"tessera": {
"command": "uvx",
"args": ["--from", "project-tessera", "tessera-mcp"]
}
}
}With pip:
{
"mcpServers": {
"tessera": {
"command": "tessera-mcp"
}
}
}Config location:
macOS:
~/Library/Application Support/Claude/claude_desktop_config.jsonWindows:
%APPDATA%\Claude\claude_desktop_config.json
Configuration
tessera setup creates workspace.yaml:
workspace:
root: /Users/you/Documents
name: my-workspace
sources:
- path: .
type: document
search:
reranker_weight: 0.7 # Semantic vs keyword balance (0.0 = keyword only, 1.0 = vector only)
max_top_k: 50 # Max results per search
ingestion:
chunk_size: 1024 # Tokens per chunk
chunk_overlap: 100 # Overlap between chunks
hooks: # Optional plugin hooks
on_memory_created:
- script: ./my-hook.shOr set TESSERA_WORKSPACE=/path/to/docs to skip config file entirely.
Environment variables:
TESSERA_API_KEY-- enable API authenticationTESSERA_VAULT_KEY-- enable AES-256 encryption for memories
Technical details
Component | Technology | Why |
Vector store | LanceDB | Embedded columnar store. No server process, handles vector + metadata queries natively |
Embeddings | fastembed/ONNX | Local inference, no API keys. |
Metadata | SQLite | File tracking, search analytics, interaction logging. Thread-safe with reentrant locks |
Memory storage | Filesystem (.md) | Human-readable, git-friendly, encryptable. YAML frontmatter for metadata |
Encryption | Pure Python AES-256-CBC | No OpenSSL dependency. PKCS7 padding, random IV per memory |
HTTP API | FastAPI | Swagger docs, Pydantic validation, async-capable |
MCP | FastMCP (stdio) | Standard MCP protocol for Claude Desktop |
Numbers
Metric | Count |
MCP tools | 58 |
HTTP endpoints | 54 |
CLI commands | 11 |
Core modules | 69 |
Lines of code | 31,000+ |
Tests | 1102 |
File types | 40+ |
License
AGPL-3.0 -- see LICENSE.
Commercial licensing: bessl.framework@gmail.com