Skip to main content
Glama

Obsidian Diary MCP Server

AI-powered journaling with local processing, automatic backlinks, and smart prompts. Combines Obsidian with Ollama for deep reflection.

Privacy: All AI processing is local via Ollama. Content never leaves your machine.

Features

  • 🧠 AI-generated reflection prompts based on recent entries

  • 🔗 Automatic [[YYYY-MM-DD]] backlinks using theme similarity

  • 🏷️ Smart #tag extraction from your writing

  • ✅ Todo extraction to organized checklists

  • 📊 Memory trace analysis with theme evolution

  • 🗓️ Sunday synthesis (weekly reflection prompts)

Requirements

  • uv, Ollama (llama3.1 or compatible model), MCP client, Obsidian vault

Setup

git clone https://github.com/madebygps/obsidian-diary-mcp.git cd obsidian-diary-mcp uv sync chmod +x start-server.sh # Configure cp .env.example .env # Edit .env: set DIARY_PATH and PLANNER_PATH (required) # Add to MCP client config (e.g., GitHub Copilot CLI) # Name: diary # Command: /full/path/to/obsidian-diary-mcp/start-server.sh

Configuration (.env):

Required: DIARY_PATH, PLANNER_PATH

Optional: OLLAMA_MODEL (default: llama3.1:latest), OLLAMA_TIMEOUT (60s), OLLAMA_TEMPERATURE (0.7), OLLAMA_NUM_PREDICT (1000 tokens)

Usage

  1. Create: "create a memory log for today" → AI-generated prompts

  2. Write: Open in Obsidian, reflect in Brain Dump section

  3. Extract: "extract todos from today's entry" → Action items to planner

  4. Link: "link today's memory log" → Auto-generates backlinks & tags

  5. Explore: Use Obsidian's backlinks panel and graph view

More Commands: "show themes from last week", "create memory trace for 30 days", "refresh memory links for 30 days"

Debugging

Logs in logs/ directory: server-YYYY-MM-DD.log (protocol), debug-YYYY-MM-DD.log (operations)

tail -f logs/debug-$(date +%Y-%m-%d).log # Watch in real-time grep ERROR logs/debug-*.log # Find errors grep "similarity" logs/debug-*.log # Debug backlinks

Troubleshooting

Server issues: Check .env exists with DIARY_PATH and PLANNER_PATH set. Run ./start-server.sh directly to test.

Ollama issues: Verify running with curl http://localhost:11434/api/tags. Pull model: ollama pull llama3.1:latest

No backlinks: Need 2+ entries with similar themes (>8% overlap). Check Brain Dump has content: grep "similarity" logs/debug-*.log

Timeouts: Increase OLLAMA_TIMEOUT (90+) and OLLAMA_NUM_PREDICT (2000+) for reasoning models.

How It Works

  • Local AI: Ollama processes entries locally—content never leaves your machine

  • Brain Dump Focus: Analyzes your writing (not prompts) for themes

  • Smart Prompts: Context-aware questions based on recent entries

  • Auto-linking: Jaccard similarity connects entries with >8% theme overlap

  • Sundays: 5 weekly synthesis prompts (vs 3 daily)

  • Todo Extraction: AI identifies action items and creates checklists

Entry Format

Each entry (YYYY-MM-DD.md):

  1. Reflection Prompts (3-5 AI-generated questions)

  2. Brain Dump (your freeform writing)

  3. Memory Links (auto-generated: [[YYYY-MM-DD]] backlinks + #tags)

License

MIT • Python 3.13+ • FastMCP 2.12.4+ • Ollama

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/madebygps/obsidian-diary-mcp'

If you have feedback or need assistance with the MCP directory API, please join our Discord server