Skip to main content
Glama

MemOS-MCP

by qinshu1109
Apache 2.0
3
  • Linux
  • Apple
simple_treekvcache_memos_config.json1.07 kB
{ "user_id": "root", "chat_model": { "backend": "huggingface", "config": { "model_name_or_path": "Qwen/Qwen3-1.7B", "temperature": 0.1, "remove_think_prefix": true, "max_tokens": 4096 } }, "mem_reader": { "backend": "simple_struct", "config": { "llm": { "backend": "ollama", "config": { "model_name_or_path": "qwen3:0.6b", "temperature": 0.8, "max_tokens": 1024, "top_p": 0.9, "top_k": 50 } }, "embedder": { "backend": "ollama", "config": { "model_name_or_path": "nomic-embed-text:latest" } }, "chunker": { "backend": "sentence", "config": { "tokenizer_or_token_counter": "gpt2", "chunk_size": 512, "chunk_overlap": 128, "min_sentences_per_chunk": 1 } } } }, "max_turns_window": 20, "top_k": 5, "enable_textual_memory": false, "enable_activation_memory": true, "enable_parametric_memory": false }

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/qinshu1109/memos-MCP'

If you have feedback or need assistance with the MCP directory API, please join our Discord server