Skip to main content
Glama

Cortex MCP

Official
by cortexapps
Dockerfile862 B
FROM python:3.13-alpine AS builder RUN apk add --no-cache \ gcc \ musl-dev \ libffi-dev \ git WORKDIR /build COPY pyproject.toml . RUN pip install --upgrade pip && \ pip wheel --no-cache-dir --wheel-dir /wheels . FROM python:3.13-alpine ENV PYTHONUNBUFFERED=1 \ PIP_NO_CACHE_DIR=1 \ # Use stdio transport for Docker as the default. Revisit after streamable-http MCP_TRANSPORT=stdio \ LOG_LEVEL=INFO \ OPENAPI_SPEC_PATH=/app/openapi.json WORKDIR /app RUN apk add --no-cache git # Copy wheels from builder and install COPY --from=builder /wheels /wheels RUN pip install --no-cache-dir /wheels/*.whl && \ rm -rf /wheels COPY src/ ./src/ COPY server.py . COPY swagger.json /app/openapi.json RUN adduser -D -u 1000 appuser && \ chown -R appuser:appuser /app USER appuser ENTRYPOINT ["python", "server.py"]

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/cortexapps/cortex-mcp'

If you have feedback or need assistance with the MCP directory API, please join our Discord server