Skip to main content
Glama

MCP Waifu Queue

by waifuai
pyproject.toml1.05 kB
[project] name = "mcp-waifu-queue" version = "0.1.0" description = "An MCP server for a conversational AI waifu character using a text queue and GPU." readme = "README.md" requires-python = ">=3.7" dependencies = [ "Flask>=2.0", "python-dotenv>=0.19", "pydantic>=2.0", "pydantic-settings>=2.0", "requests>=2.25", "gunicorn>=20.1", "anyio>=4.3", "redis>=4.0", "rq>=1.10", "mcp>=1.1.0" ] [project.optional-dependencies] test = [ "pytest>=7.0", "pytest-cov>=4.0", "pytest-asyncio>=0.18", "fakeredis[lua]>=2.0", "requests-mock>=1.9" ] [build-system] requires = ["setuptools>=42", "wheel"] build-backend = "setuptools.build_meta" [tool.setuptools] packages = ["mcp_waifu_queue"] [tool.pytest.ini_options] pythonpath = [ ".", "mcp_waifu_queue" ] testpaths = ["tests"] addopts = "--cov=mcp_waifu_queue --cov-report term-missing" [tool.ruff] line-length = 88 target-version = "py310" [tool.ruff.lint] select = ["E", "F", "I", "W", "C", "N", "B", "SIM", "TCH", "RUF"] ignore = ["E501"]

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/waifuai/mcp-waifu-queue'

If you have feedback or need assistance with the MCP directory API, please join our Discord server