Skip to main content
Glama

English | 中文

One Memory. All CLIs. Never Compacted. Exact Search.

Session history management for AI coding assistants. Never lose your conversations again.

Features

  • On-demand search - You control when to search; automatic injection is opt-in

  • Original preservation - Raw messages always kept; summaries are optional layers

  • Multi-CLI support - Claude Code, Codex, OpenCode, Gemini in one database

  • Powerful search - Full-text (FTS5) + semantic vectors + hybrid ranking

  • MCP integration - Search directly from your AI CLI

  • REST API - Integrate into any workflow

  • Local storage - All data stays on your machine

Quick Start

Full

brew install vimo-ai/tap/memex

# Verify server is running
curl http://localhost:10013/health

Lite

Zero-dependency CLI, reads local session data directly:

brew install vimo-ai/tap/memex-lite

memex search "anything you want"
memex list -n 10

Docker

macOS / Linux:

docker run -d -p 10013:10013 \
  -v ~/.vimo:/data \
  -v ~/.claude/projects:/claude:ro \
  -v ~/.codex:/codex:ro \                              # 可选: Codex
  -v ~/.local/share/opencode:/opencode:ro \            # 可选: OpenCode
  -v ~/.gemini/tmp:/gemini:ro \                        # 可选: Gemini
  -e OLLAMA_HOST=http://host.docker.internal:11434 \   # 可选: 本机 Ollama (Docker Desktop)
  ghcr.io/vimo-ai/memex:latest

Windows (PowerShell):

docker run -d -p 10013:10013 `
  -v "$env:USERPROFILE\.vimo:/data" `
  -v "$env:USERPROFILE\.claude\projects:/claude:ro" `
  -v "$env:USERPROFILE\.codex:/codex:ro" `             # 可选: Codex
  -v "$env:LOCALAPPDATA\opencode:/opencode:ro" `       # 可选: OpenCode
  -v "$env:USERPROFILE\.gemini\tmp:/gemini:ro" `       # 可选: Gemini
  -e OLLAMA_HOST=http://host.docker.internal:11434 `   # 可选: 本机 Ollama (Docker Desktop)
  ghcr.io/vimo-ai/memex:latest

Binary downloads available at Releases.

Configure MCP

# Claude Code
claude mcp add memex -- npx -y mcp-remote http://localhost:10013/api/mcp

# Codex
codex mcp add memex -- npx -y mcp-remote http://localhost:10013/api/mcp

# Gemini
gemini mcp add --transport http memex http://localhost:10013/api/mcp

# OpenCode - edit ~/.config/opencode/opencode.json
# { "mcp": { "memex": { "type": "remote", "url": "http://localhost:10013/api/mcp" } } }

Then search in your AI CLI:

use memex search "anything you want"

Hooks (Optional)

Auto-inject relevant memory context into Claude Code sessions. See Hook Documentation for setup.

Documentation

https://vimoai.dev/docs/memex

Community

Discord

Join our Discord server for discussions, support, and updates.

-
security - not tested
A
license - permissive license
-
quality - not tested

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/vimo-ai/memex'

If you have feedback or need assistance with the MCP directory API, please join our Discord server