MemGPT MCP Server

# MemGPT MCP Server A TypeScript-based MCP server that implements a memory system for LLMs. It provides tools for chatting with different LLM providers while maintaining conversation history. ## Features ### Tools - `chat` - Send a message to the current LLM provider - Takes a message parameter - Supports multiple providers (OpenAI, Anthropic, OpenRouter, Ollama) - `get_memory` - Retrieve conversation history - Optional `limit` parameter to specify number of memories to retrieve - Pass `limit: null` for unlimited memory retrieval - Returns memories in chronological order with timestamps - `clear_memory` - Clear conversation history - Removes all stored memories - `use_provider` - Switch between different LLM providers - Supports OpenAI, Anthropic, OpenRouter, and Ollama - Persists provider selection - `use_model` - Switch to a different model for the current provider - Supports provider-specific models: - Anthropic Claude Models: - Claude 3 Series: - `claude-3-haiku`: Fastest response times, ideal for tasks like customer support and content moderation - `claude-3-sonnet`: Balanced performance for general-purpose use - `claude-3-opus`: Advanced model for complex reasoning and high-performance tasks - Claude 3.5 Series: - `claude-3.5-haiku`: Enhanced speed and cost-effectiveness - `claude-3.5-sonnet`: Superior performance with computer interaction capabilities - OpenAI: 'gpt-4o', 'gpt-4o-mini', 'gpt-4-turbo' - OpenRouter: Any model in 'provider/model' format (e.g., 'openai/gpt-4', 'anthropic/claude-2') - Ollama: Any locally available model (e.g., 'llama2', 'codellama') - Persists model selection ## Development Install dependencies: ```bash npm install ``` Build the server: ```bash npm run build ``` For development with auto-rebuild: ```bash npm run watch ``` ## Installation To use with Claude Desktop, add the server config: On MacOS: `~/Library/Application Support/Claude/claude_desktop_config.json` On Windows: `%APPDATA%/Claude/claude_desktop_config.json` ```json { "mcpServers": { "letta-memgpt": { "command": "/path/to/memgpt-server/build/index.js", "env": { "OPENAI_API_KEY": "your-openai-key", "ANTHROPIC_API_KEY": "your-anthropic-key", "OPENROUTER_API_KEY": "your-openrouter-key" } } } } ``` ### Environment Variables - `OPENAI_API_KEY` - Your OpenAI API key - `ANTHROPIC_API_KEY` - Your Anthropic API key - `OPENROUTER_API_KEY` - Your OpenRouter API key ### Debugging Since MCP servers communicate over stdio, debugging can be challenging. We recommend using the [MCP Inspector](https://github.com/modelcontextprotocol/inspector): ```bash npm run inspector ``` The Inspector will provide a URL to access debugging tools in your browser. ## Recent Updates ### Claude 3 and 3.5 Series Support (March 2024) - Added support for latest Claude models: - Claude 3 Series (Haiku, Sonnet, Opus) - Claude 3.5 Series (Haiku, Sonnet) ### Unlimited Memory Retrieval - Added support for retrieving unlimited conversation history - Use `{ "limit": null }` with the `get_memory` tool to retrieve all stored memories - Use `{ "limit": n }` to retrieve the n most recent memories - Default limit is 10 if not specified