Skip to main content
Glama

Comfy MCP Server

by lalanikarim
pyproject.toml816 B
[project] name = "comfy-mcp-server" version = "0.1.11" description = "MCP Server for ComfyUI text to image Workflow" readme = "README.md" requires-python = ">=3.10" dependencies = [ "langchain>=0.3.19", "langchain-ollama>=0.2.3", "mcp[cli]>=1.2.1", ] authors = [ {name = "Karim Lalani", email = "jimmy00784@gmail.com"}, ] maintainers = [ {name = "Karim Lalani", email = "jimmy00784@gmail.com"}, ] license = "MIT" license-files = ["LICENSE"] [project.urls] Homepage = "https://github.com/lalanikarim/comfy-mcp-server/" Repository = "https://github.com/lalanikarim/comfy-mcp-server/" [project.scripts] comfy-mcp-server = "comfy_mcp_server:run_server" [build-system] requires = ["hatchling"] build-backend = "hatchling.build" [tool.hatch.build.targets.wheel] packages = ["src/comfy_mcp_server"]

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/lalanikarim/comfy-mcp-server'

If you have feedback or need assistance with the MCP directory API, please join our Discord server