Skip to main content
Glama

HK Prompt MCP Server

by hkopenai
Dockerfile711 B
# Multi-stage build for a slimmer image # Stage 1: Build dependencies FROM python:3.13-slim AS builder WORKDIR /app COPY pyproject.toml . COPY hkopenai/ ./hkopenai/ RUN pip install --user --no-cache-dir . # Stage 2: Runtime image FROM python:3.13-slim WORKDIR /app # Copy only the necessary files from the builder stage COPY --from=builder /root/.local /root/.local # Copy the application code COPY hkopenai/ ./hkopenai/ # COPY LICENSE . # COPY README.md . # Set PATH to include user-installed packages ENV PATH=/root/.local/bin:$PATH # Expose the port the app runs on EXPOSE 8000 # Command to run the MCP server in SSE CMD ["python", "-m", "hkopenai.diagram_as_code_mcp_server", "--sse", "--host", "0.0.0.0"]

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/hkopenai/document_as_code'

If you have feedback or need assistance with the MCP directory API, please join our Discord server