Skip to main content
Glama

MCP Memory Service

# Core dependencies chromadb==0.5.23 tokenizers==0.20.3 websockets>=11.0.3 mcp>=1.0.0,<2.0.0 # Platform-specific recommendations (DO NOT UNCOMMENT - handled by install.py) # For macOS with Intel CPUs: # torch==2.0.1 # torchvision==2.0.1 # torchaudio==2.0.1 # sentence-transformers==2.2.2 # For macOS with Apple Silicon: # torch>=2.0.0 # torchvision>=0.15.0 # torchaudio>=2.0.0 # sentence-transformers>=2.2.2 # For Windows with NVIDIA GPU: # pip install torch==2.1.0 torchvision==2.1.0 torchaudio==2.1.0 --index-url=https://download.pytorch.org/whl/cu118 # sentence-transformers>=2.2.2 # For Windows with DirectML: # torch==2.1.0 # torchvision==2.1.0 # torchaudio==2.1.0 # torch-directml>=0.2.0 # sentence-transformers>=2.2.2 # For Linux with CUDA: # torch>=2.0.0 # torchvision>=0.15.0 # torchaudio>=2.0.0 # sentence-transformers>=2.2.2 # For CPU-only fallback (all platforms): # torch==1.13.1 # torchvision==0.14.1 # torchaudio==0.13.1 # sentence-transformers==2.2.2 # onnxruntime>=1.15.0 # Note: PyTorch and sentence-transformers will be installed by install.py with # appropriate platform-specific versions. Do not install manually unless you # encounter issues with the automatic installation.

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/doobidoo/mcp-memory-service'

If you have feedback or need assistance with the MCP directory API, please join our Discord server