Skip to main content
Glama

Qiniu MCP Server

Official
by qiniu
pyproject.toml629 B
[project] name = "qiniu-mcp-server" version = "1.2.3" description = "A MCP server project of Qiniu." requires-python = ">=3.12" authors = [ { name = "Qiniu", email = "sdk@qiniu.com" }, ] keywords = ["qiniu", "mcp", "llm"] dependencies = [ "aioboto3>=13.2.0", "fastjsonschema>=2.21.1", "httpx>=0.28.1", "mcp[cli]>=1.0.0", "openai>=1.66.3", "pip>=25.0.1", "python-dotenv>=1.0.1", "qiniu>=7.16.0", ] [build-system] requires = [ "hatchling",] build-backend = "hatchling.build" [project.scripts] qiniu-mcp-server = "mcp_server:main" [tool.hatch.build.targets.wheel] packages = ["src/mcp_server"]

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/qiniu/qiniu-mcp-server'

If you have feedback or need assistance with the MCP directory API, please join our Discord server