Skip to main content
Glama

AWS Model Context Protocol Server

by alexei-led
services: aws-mcp-server: # Use either local build or official image from GitHub Packages build: context: ../../ dockerfile: ./deploy/docker/Dockerfile # Alternatively, use the pre-built multi-arch image # image: ghcr.io/alexei-led/aws-mcp-server:latest ports: - "8000:8000" volumes: - ~/.aws://home/appuser/.aws:ro # Mount AWS credentials as read-only environment: - AWS_PROFILE=default # Specify default AWS profile - AWS_MCP_TIMEOUT=300 # Default timeout in seconds (5 minutes) - AWS_MCP_TRANSPORT=stdio # Transport protocol ("stdio" or "sse") # - AWS_MCP_MAX_OUTPUT=100000 # Uncomment to set max output size restart: unless-stopped # To build multi-architecture images: # 1. Set up Docker buildx: docker buildx create --name mybuilder --use # 2. Build and push the multi-arch image: # docker buildx build --platform linux/amd64,linux/arm64 -t yourrepo/aws-mcp-server:latest --push .

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/alexei-led/aws-mcp-server'

If you have feedback or need assistance with the MCP directory API, please join our Discord server