Skip to main content
Glama
README.md2.44 kB
# ComfyUI_MCP This repository provides a Model Context Protocol (MCP) server that exposes local ComfyUI workflow files and the remote ComfyUI model catalog to compatible LLM clients. ## Features - Discovers workflow `.json` files stored under the `workflows/` directory and exposes them as MCP resources. - Tools for listing workflow metadata and reading workflow contents. - Tool for recursively querying the ComfyUI `/api/models` endpoints so an LLM can inspect the available checkpoints, LoRAs, and other assets. - Configuration via CLI arguments or environment variables so the server can be launched from JSON descriptors (e.g., Cursor MCP definitions). ## Installation Create a virtual environment (recommended) and install the package in editable mode with your preferred installer: ```bash pip install -e . ``` The project also works with [uv](https://github.com/astral-sh/uv) so you can install or run it without using `pip` directly: ```bash uv pip install -e . # or run without installing into the current environment uvx --from . comfyui-mcp --help ``` ## Running the server The installed `comfyui-mcp` entry point launches the MCP server over stdio. Common configuration options can be supplied either as CLI flags or environment variables: | Purpose | CLI flag | Environment variable | Default | |---------|----------|----------------------|---------| | Workflow directory | `--workflow-dir` | `COMFYUI_WORKFLOW_DIR` | `<repo>/workflows` | | ComfyUI API base URL | `--api-base-url` | `COMFYUI_API_BASE_URL` | `http://127.0.0.1:8188/` | | HTTP timeout (seconds) | `--http-timeout` | `COMFYUI_HTTP_TIMEOUT` | `30` | | Log level | `--log-level` | `COMFYUI_MCP_LOG_LEVEL` | `INFO` | Example stdio launch configuration for a Cursor MCP JSON definition: If you prefer `uvx`, the same configuration can be expressed as: ```json { "comfyui": { "command": "uvx", "args": [ "--from", ".", "comfyui-mcp", "--workflow-dir", "./workflows", "--api-base-url", "http://127.0.0.1:8188/" ] } } ``` When referencing the project locally with `uvx`, ensure the working directory is set to the repository root (or adjust the `--from` path accordingly) so the package can be resolved without requiring it to be published to an external index. Place your ComfyUI workflow files in the `workflows/` directory (or whatever directory you configure) so they are available to the LLM.

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/neutrinotek/ComfyUI_MCP'

If you have feedback or need assistance with the MCP directory API, please join our Discord server