Skip to main content
Glama

Sequential Questioning MCP Server

by bitgeese
pyproject.toml1.7 kB
[build-system] requires = ["hatchling"] build-backend = "hatchling.build" [project] name = "sequential-questioning-mcp-server" version = "1.0.0" description = "A Sequential Questioning MCP Server for facilitating multi-round questioning to gather specific, contextual information from end-users" authors = [ {name = "Project Developer"} ] readme = "README.md" requires-python = ">=3.10" dependencies = [ "fastapi>=0.109.0", "uvicorn[standard]>=0.27.0", "fastapi-mcp>=0.3.0", "langchain>=0.1.0", "langchain-openai>=0.0.2", "langchain-community>=0.0.10", "openai>=1.10.0", "qdrant-client>=1.7.0", "sqlalchemy>=2.0.0", "alembic>=1.13.0", "pydantic>=2.5.0", "pydantic-settings>=2.1.0", "httpx>=0.26.0", "python-dotenv>=1.0.0", "asyncpg>=0.29.0", "psycopg2-binary>=2.9.9", "aiosqlite>=0.19.0", "numpy>=1.25.0", "tenacity>=8.2.0", ] [project.optional-dependencies] dev = [ "pytest>=7.4.0", "pytest-asyncio>=0.21.1", "black>=23.7.0", "isort>=5.12.0", "mypy>=1.5.0", "ruff>=0.1.0", ] [tool.hatch.build.targets.wheel] packages = ["app"] [tool.pytest.ini_options] testpaths = ["tests"] python_files = "test_*.py" [tool.black] line-length = 88 target-version = ["py310"] [tool.isort] profile = "black" line_length = 88 [tool.mypy] python_version = "3.10" disallow_untyped_defs = true disallow_incomplete_defs = true check_untyped_defs = true disallow_untyped_decorators = true no_implicit_optional = true strict_optional = true warn_redundant_casts = true warn_return_any = true warn_unused_ignores = true [tool.ruff] line-length = 88 target-version = "py310" select = ["E", "F", "I"] ignore = []

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/bitgeese/sequential-questioning'

If you have feedback or need assistance with the MCP directory API, please join our Discord server