Skip to main content
Glama

code2prompt-mcp

by ODAncona
pyproject.toml630 B
[project] name = "code2prompt-mcp" version = "0.1.0" description = "MCP server for Code2Prompt" authors = [ { name = "Olivier D'Ancona", email = "olivier.dancona@master.hes-so.ch" }, ] dependencies = [ "mcp>=1.4.1", "httpx>=0.28.1", "dotenv>=0.9.9", "colorlog>=6.9.0", "code2prompt-rs>=3.2.1", ] readme = "README.md" requires-python = ">= 3.8" [build-system] requires = ["hatchling"] build-backend = "hatchling.build" [tool.rye] managed = true dev-dependencies = [] virtual = true [tool.hatch.metadata] allow-direct-references = true [tool.hatch.build.targets.wheel] packages = ["src/code2prompt_mcp"]

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/ODAncona/code2prompt-mcp'

If you have feedback or need assistance with the MCP directory API, please join our Discord server