Search for:
Why this server?
A Model Context Protocol server for Claude Desktop that provides structured memory management across chat sessions, allowing Claude to maintain context and build a knowledge base within project directories.
Why this server?
A TypeScript-based server that provides a memory system for Large Language Models (LLMs), allowing users to interact with multiple LLM providers while maintaining conversation history and offering tools for managing providers and model configurations.
Why this server?
Enables LLMs to search, retrieve, and manage documents through Rememberizer's knowledge management API.
Why this server?
Provides a centralized MCP-based system for managing and accessing multi-project memory banks remotely, with features like project isolation, file structure validation, and type-safe operations.
Why this server?
Memory manager for AI apps and Agents using various graph and vector stores and allowing ingestion from 30+ data sources
Why this server?
A server for managing project documentation and context across Claude AI sessions through global and branch-specific memory banks, enabling consistent knowledge management with structured JSON document storage.
Why this server?
A server that enables running a Model Context Protocol in a web browser with functionalities including arithmetic operations and session-based key-value storage.
Why this server?
Provides knowledge graph functionality for managing entities, relations, and observations in memory with strict validation rules to maintain data consistency.
Why this server?
Facilitates initiating Ollama queries via Claude and manages a simple note storage system with capabilities to add, summarize, and access notes using custom URIs.
Why this server?
A flexible memory system for AI applications that supports multiple LLM providers and can be used either as an MCP server or as a direct library integration, enabling autonomous memory management without explicit commands.