Supports academic paper citation and publication, with references to arXiv papers related to memory systems for LLMs.
Provides community support through Discord server integration.
Enables communication with GitHub through issues, discussions, and pull requests for bug reporting and community support.
Supports integration with Ollama for local model deployment and inference.
Includes benchmarking against OpenAI models and potential integration capabilities.
Provides integration with PyTorch for model acceleration and GPU support.
Offers community support and integration through WeChat groups.
MemOS - MCP Integration
MemOS记忆系统的MCP(Model Context Protocol)集成版本,专为个人AI助手场景优化。
基于原始MemOS项目,针对Claude Desktop和MCP协议进行了深度优化。
🚀 特性
核心功能
- 智能记忆管理: 基于向量数据库的语义记忆存储和检索
- MCP协议支持: 完整的Model Context Protocol集成
- 多模型支持: 支持Qwen3、SiliconFlow等多种嵌入和重排模型
- 时间感知检索: 基于时间衰减的智能记忆排序
- 缓存优化: LRU缓存系统,显著提升检索性能
技术架构
- 分层设计: 表示层、业务逻辑层、数据访问层的清晰分离
- MVP管理器: 统一的记忆管理接口,支持增强版和基础版自动切换
- 容量管理: 智能的记忆容量管理和自动压缩
- 并发控制: 多进程安全的并发访问控制
- Website: https://memos.openmem.net/
- Documentation: https://memos-docs.openmem.net/home/overview/
- API Reference: https://memos-docs.openmem.net/docs/api/info/
- Source Code: https://github.com/MemTensor/MemOS
📈 Performance Benchmark
MemOS demonstrates significant improvements over baseline memory solutions in multiple reasoning tasks.
Model | Avg. Score | Multi-Hop | Open Domain | Single-Hop | Temporal Reasoning |
---|---|---|---|---|---|
OpenAI | 0.5275 | 0.6028 | 0.3299 | 0.6183 | 0.2825 |
MemOS | 0.7331 | 0.6430 | 0.5521 | 0.7844 | 0.7321 |
Improvement | +38.98% | +6.67% | +67.35% | +26.86% | +159.15% |
💡 Temporal reasoning accuracy improved by 159% compared to the OpenAI baseline.
Details of End-to-End Evaluation on LOCOMO
Note
Comparison of LLM Judge Scores across five major tasks in the LOCOMO benchmark. Each bar shows the mean evaluation score judged by LLMs for a given method-task pair, with standard deviation as error bars. MemOS-0630 consistently outperforms baseline methods (LangMem, Zep, OpenAI, Mem0) across all task types, especially in multi-hop and temporal reasoning scenarios.
✨ Key Features
- 🧠 Memory-Augmented Generation (MAG): Provides a unified API for memory operations, integrating with LLMs to enhance chat and reasoning with contextual memory retrieval.
- 📦 Modular Memory Architecture (MemCube): A flexible and modular architecture that allows for easy integration and management of different memory types.
- 💾 Multiple Memory Types:
- Textual Memory: For storing and retrieving unstructured or structured text knowledge.
- Activation Memory: Caches key-value pairs (
KVCacheMemory
) to accelerate LLM inference and context reuse. - Parametric Memory: Stores model adaptation parameters (e.g., LoRA weights).
- 🔌 Extensible: Easily extend and customize memory modules, data sources, and LLM integrations.
🚀 Getting Started
Here's a quick example of how to create a MemCube
, load it from a directory, access its memories, and save it.
What about MOS
(Memory Operating System)? It's a higher-level orchestration layer that manages multiple MemCubes and provides a unified API for memory operations. Here's a quick example of how to use MOS:
For more detailed examples, please check out the examples
directory.
📦 Installation
Warning
MemOS is compatible with Linux, Windows, and macOS.
However, if you're using macOS, please note that there may be dependency issues that are difficult to resolve.
For example, compatibility with macOS 13 Ventura is currently challenging.
Install via pip
Development Install
To contribute to MemOS, clone the repository and install it in editable mode:
Optional Dependencies
Ollama Support
To use MemOS with Ollama, first install the Ollama CLI:
Transformers Support
To use functionalities based on the transformers
library, ensure you have PyTorch installed (CUDA version recommended for GPU acceleration).
💬 Community & Support
Join our community to ask questions, share your projects, and connect with other developers.
- GitHub Issues: Report bugs or request features in our GitHub Issues.
- GitHub Pull Requests: Contribute code improvements via Pull Requests.
- GitHub Discussions: Participate in our GitHub Discussions to ask questions or share ideas.
- Discord: Join our Discord Server.
- WeChat: Scan the QR code to join our WeChat group.
📜 Citation
Note
We publicly released the Short Version on May 28, 2025, making it the earliest work to propose the concept of a Memory Operating System for LLMs.
If you use MemOS in your research, we would appreciate citations to our papers.
🙌 Contributing
We welcome contributions from the community! Please read our contribution guidelines to get started.
📄 License
MemOS is licensed under the Apache 2.0 License.
📰 News
Stay up to date with the latest MemOS announcements, releases, and community highlights.
- 2025-07-07 – 🎉 MemOS 1.0 (Stellar) Preview Release: A SOTA Memory OS for LLMs is now open-sourced.
- 2025-07-04 – 🎉 MemOS Paper Released: MemOS: A Memory OS for AI System was published on arXiv.
- 2025-05-28 – 🎉 Short Paper Uploaded: MemOS: An Operating System for Memory-Augmented Generation (MAG) in Large Language Models was published on arXiv.
- 2024-07-04 – 🎉 Memory3 Model Released at WAIC 2024: The new memory-layered architecture model was unveiled at the 2024 World Artificial Intelligence Conference.
- 2024-07-01 – 🎉 Memory3 Paper Released: Memory3: Language Modeling with Explicit Memory introduces the new approach to structured memory in LLMs.
This server cannot be installed
A Model Context Protocol integration for the MemOS memory system, optimized for personal AI assistant scenarios with intelligent memory management and retrieval capabilities.
Related MCP Servers
- -securityFlicense-qualityA comprehensive Model Context Protocol server implementation that enables AI assistants to interact with file systems, databases, GitHub repositories, web resources, and system tools while maintaining security and control.Last updated -161TypeScript
- AsecurityFlicenseAqualityA Model Context Protocol server that enables AI assistants to interact with the Omi API for retrieving and creating conversations and memories for users.Last updated -42TypeScript
- -securityAlicense-qualityA Model Context Protocol server that provides AI agents with persistent memory capabilities through Mem0, allowing them to store, retrieve, and semantically search memories.Last updated -477PythonMIT License
- -securityAlicense-qualityA Model Context Protocol server that integrates AI assistants with Mem0.ai's persistent memory system, allowing models to store, retrieve, search, and manage different types of memories.Last updated -9PythonMIT License