Skip to main content
Glama

MCP LLMS-TXT Documentation Server

by teddylee777
pyproject.toml616 B
[project] name = "mcpdoc" version = "0.0.4" description = "Server llms-txt documentation over MCP" readme = "README.md" requires-python = ">=3.10" dependencies = [ "httpx>=0.28.1", "markdownify>=1.1.0", "mcp[cli]>=1.4.1", "pyyaml>=6.0.1", "requests>=2.32.3", ] [project.scripts] mcpdoc = "mcpdoc.cli:main" [dependency-groups] test = [ "pytest>=8.3.4", "pytest-asyncio>=0.25.3", "pytest-cov>=6.0.0", "pytest-mock>=3.14.0", "pytest-socket>=0.7.0", "pytest-timeout>=2.3.1", "ruff>=0.9.7", ] [build-system] requires = ["hatchling"] build-backend = "hatchling.build"

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/teddylee777/mcpdoc'

If you have feedback or need assistance with the MCP directory API, please join our Discord server