Integrates with OpenAI's embedding models to generate high-quality semantic vector representations for stored decisions and learnings.
Supports PostgreSQL with the pgvector extension as a scalable backend for storing and performing similarity searches on agent memories and project context.
Provides a local SQLite-based storage system for persistent vector memory, allowing AI agents to save and retrieve semantic context across sessions without external dependencies.
Click on "Install Server".
Wait a few minutes for the server to deploy. Once ready, it will show a "Started" state.
In the chat, type
@followed by the MCP server name and your instructions, e.g., "@MCP Vector Memorysearch for the architectural decisions we made about the database"
That's it! The server will respond to your query, and you can continue using it as needed.
Here is a step-by-step guide with screenshots.
MCP Vector Memory
Persistent vector memory for AI coding agents. Give your AI agents long-term memory that survives across conversations.
Works with Antigravity · VS Code / Copilot · Claude Code · Codex
The Problem
AI coding agents forget everything between conversations. Every new session starts from zero — they repeat questions, lose architectural decisions, and forget how your project is set up.
The Solution
MCP Vector Memory gives your agents persistent, semantic memory. Agents can save and search memories using natural language. Memories are stored locally in SQLite with vector similarity search — no API keys, no cloud, no cost.
Quick Start
1. Install
2. Configure Your IDE
Add to ~/.gemini/antigravity/mcp_config.json:
Create .vscode/mcp.json in your workspace:
Add to ~/.claude/claude_desktop_config.json:
Add to your Codex MCP configuration:
3. Restart your IDE
That's it. Your agents now have persistent memory.
Tools Available
Tool | Description |
| Semantic search over past memories |
| Save a decision, context, or learning |
| List all projects with memory counts |
| Memory system statistics |
How It Works
Agent calls
save_memory→ text is embedded locally → stored in SQLite with vector indexAgent calls
search_memory→ query is embedded → SQLite finds most similar memoriesEverything runs locally. No API calls, no cloud, no cost.
Configuration
All configuration is via environment variables:
Variable | Default | Description |
|
| Database backend ( |
|
| SQLite data directory |
|
| Embedding provider ( |
|
| Embedding model name |
PostgreSQL Backend (Optional)
For production deployments with PostgreSQL + pgvector:
Docker (PostgreSQL)
See docker-compose.yml for the full setup.
OpenAI Embeddings (Optional)
For higher quality embeddings via OpenAI API:
Contributing
Contributions are welcome! Please see CONTRIBUTING.md for guidelines.
License
MIT — see LICENSE