LLM Wiki Kit
Provides tools for OpenAI Codex to process and ingest raw documents, write and update structured wiki pages, query the knowledge base via search, and manage persistent markdown documentation with cross-references.
Click on "Install Server".
Wait a few minutes for the server to deploy. Once ready, it will show a "Started" state.
In the chat, type
@followed by the MCP server name and your instructions, e.g., "@LLM Wiki Kitingest the papers from raw/ and update the knowledge base"
That's it! The server will respond to your query, and you can continue using it as needed.
Here is a step-by-step guide with screenshots.
π llm-wiki-kit
An MCP server that implements Karpathy's LLM Wiki pattern - persistent, LLM-maintained knowledge bases that compound over time.
Instead of RAG (rediscovering knowledge from scratch on every query), the LLM incrementally builds and maintains a structured wiki with interlinked markdown files, cross-references, summaries, and synthesis that get richer with every source you add.
Why?
The tedious part of maintaining a knowledge base isn't the reading or thinking, it's the bookkeeping. Updating cross-references, keeping summaries current, noting contradictions, maintaining consistency. LLMs are perfect for this. You curate and direct. The LLM does everything else.
Example use case: The Research Loop
Imagine you are researching a new and complex technology like LLM speculative decoding. Instead of reading 10 papers and taking manual notes, you use llm-wiki-kit to let your agent build a state map over time.
The Workflow
Human: drops 3 PDFs into
raw/Human: "Analyze these papers and update the KB. Pay special attention to KV cache optimizations."
Agent (via MCP):
Calls
wiki_ingestfor each paperCalls
wiki_write_pageto createconcepts/speculative_decoding.mdCalls
wiki_write_pageto updatesynthesis/cache_strategies.mdand link it to the papersCalls
wiki_lintto ensure the new "Draft Model" concept is cross-referenced with existing "Inference" pages
The Result
Two weeks later, you start a fresh chat session in Cursor or Claude Code. You do not need to re-upload the papers or re-explain what you learned. You ask:
"Based on our research so far, which draft model architecture is most efficient for Llama 3?"
Your agent calls wiki_search, reads the synthesis pages it wrote earlier, and answers from accumulated evidence:
"Based on the compiled evidence in your KB, the Eagle architecture is currently leading because..."
Quick Start
Install
Requires Python 3.10+.
# With uv (recommended)
uv pip install git+https://github.com/iamsashank09/llm-wiki-kit.git
# With pip
pip install git+https://github.com/iamsashank09/llm-wiki-kit.gitNote: PyPI publishing is coming soon. For now, install directly from GitHub.
Initialize a wiki
mkdir my-research && cd my-research
llm-wiki-kit init --agent claudeThis creates:
my-research/
βββ raw/ β Drop source files here (immutable)
βββ wiki/ β LLM-maintained wiki pages
β βββ index.md β Master index
β βββ log.md β Chronological operation log
βββ WIKI.md β Schema file for your LLM agentConnect to your LLM agent
Claude Desktop / Claude Code
Add to your MCP config (claude_desktop_config.json):
{
"mcpServers": {
"llm-wiki-kit": {
"command": "llm-wiki-kit",
"args": ["serve", "--root", "/path/to/my-research"],
"env": {}
}
}
}OpenAI Codex
Add the server with:
codex mcp add llm-wiki-kit -- llm-wiki-kit serve --root /path/to/my-researchOr add it manually to ~/.codex/config.toml:
[mcp_servers.llm-wiki-kit]
command = "llm-wiki-kit"
args = ["serve", "--root", "/path/to/my-research"]Cursor
Add to .cursor/mcp.json:
{
"mcpServers": {
"llm-wiki-kit": {
"command": "llm-wiki-kit",
"args": ["serve", "--root", "/path/to/my-research"]
}
}
}Windsurf
Add to ~/.codeium/windsurf/mcp_config.json:
{
"mcpServers": {
"llm-wiki-kit": {
"command": "llm-wiki-kit",
"args": ["serve", "--root", "/path/to/my-research"]
}
}
}Any MCP-compatible agent
llm-wiki-kit serve --root /path/to/my-researchThe server uses stdio transport and is compatible with any MCP client.
Use it
Once connected, your LLM agent has these tools:
Tool | Description |
| Initialize a new wiki |
| Process a source document |
| Create/update a wiki page |
| Read a wiki page |
| Full-text search (FTS5) |
| Health-check for issues |
| Overview of wiki state |
| Append to operation log |
Example workflow:
Drop an article into
raw/Tell your agent: "Ingest raw/article.md"
The agent reads it, creates wiki pages, cross-references related concepts, updates the index
Ask questions: "How does X relate to Y?" and the agent searches the wiki and synthesizes
Periodically: "Run a lint pass" to catch broken links, orphan pages, and contradictions
Architecture
Three layers, per Karpathy's design:
βββββββββββββββββββββββββββββββββββββββββββββββ
β You (the human) β
β Source, direct, ask questions, think β
ββββββββββββββββββββ¬βββββββββββββββββββββββββββ
β
ββββββββββββββββββββΌβββββββββββββββββββββββββββ
β Wiki (wiki/) β
β LLM-maintained markdown files β
β Summaries, entities, concepts, synthesis β
β Cross-referenced with [[wiki links]] β
ββββββββββββββββββββ¬βββββββββββββββββββββββββββ
β
ββββββββββββββββββββΌβββββββββββββββββββββββββββ
β Raw Sources (raw/) β
β Articles, papers, notes (immutable) β
βββββββββββββββββββββββββββββββββββββββββββββββSearch
Wiki pages are indexed using SQLite FTS5 with Porter stemming, with zero external dependencies and instant setup. Search supports:
Simple queries:
"attention mechanism"Boolean:
transformer AND attentionNegation:
transformer NOT GPTPrefix:
trans*
Lint
The lint tool checks for:
Broken links:
[[Page Name]]references to pages that don't existOrphan pages: pages with no inbound links
Empty pages: pages with only a title or no content
Use Cases
Technical onboarding: Ingest an entire codebase's documentation into a wiki so an agent can answer architecture questions quickly
Project state: Maintain a project wiki where the agent tracks current bugs, architectural decisions, and TODOs across multiple chat sessions
Competitive intel: Feed the agent market reports and let it maintain a living landscape wiki that updates as new data arrives
Tips
Transparency and auditing: Since the wiki is just a folder of markdown files, you can point any viewer like VS Code, Zed, or Obsidian at the directory to audit the agent's work or visualize its internal knowledge graph. No GUI is required for the agent to function
The wiki is just markdown files in a git repo, so version history and collaboration come free
Start small. Even 5-10 sources produce a surprisingly useful wiki
Let the LLM cross-reference aggressively because the value is in the connections
Development
git clone https://github.com/iamsashank09/llm-wiki-kit
cd llm-wiki-kit
uv venv && source .venv/bin/activate
uv pip install -e .Credits
Based on the LLM Wiki idea by Andrej Karpathy.
License
MIT
This server cannot be installed
Resources
Unclaimed servers have limited discoverability.
Looking for Admin?
If you are the server author, to access and configure the admin panel.
Latest Blog Posts
MCP directory API
We provide all the information about MCP servers via our MCP API.
curl -X GET 'https://glama.ai/api/mcp/v1/servers/iamsashank09/llm-wiki-kit'
If you have feedback or need assistance with the MCP directory API, please join our Discord server