Skip to main content
Glama

FastMCP Boilerplate

by rainer85ah
docker-compose.yaml1.28 kB
networks: ai-net: driver: bridge enable_ipv6: false services: ollama: image: ollama/ollama:0.9.0 container_name: ollama ports: - "0.0.0.0:11434:11434" volumes: - ${MODELS_LOCATION}:/root/.ollama networks: - ai-net environment: OLLAMA_NUM_THREADS: "20" cap_add: - IPC_LOCK logging: driver: "json-file" options: max-size: "10m" max-file: "3" openweb: image: ghcr.io/open-webui/open-webui:v0.6.13 container_name: openweb depends_on: - ollama - mcp ports: - "0.0.0.0:3000:8080" volumes: - ${OPENWEB_LOCATION}:/app/backend/data environment: - OLLAMA_BASE_URL=${OLLAMA_BASE_URL} - WEBUI_SECRET_KEY=${WEBUI_SECRET_KEY} networks: - ai-net logging: driver: "json-file" options: max-size: "10m" max-file: "3" mcp: image: ghcr.io/rainer85ah/mcp-server:latest container_name: mcp depends_on: - ollama ports: - "0.0.0.0:8000:8000" environment: - OLLAMA_BASE_URL=${OLLAMA_BASE_URL} - DEFAULT_MODEL=${DEFAULT_MODEL} networks: - ai-net logging: driver: "json-file" options: max-size: "10m" max-file: "3"

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/rainer85ah/mcp-server'

If you have feedback or need assistance with the MCP directory API, please join our Discord server