Skip to main content
Glama

cognee-mcp

pyproject.toml1.09 kB
[project] name = "cognee-mcp" version = "0.4.0" description = "Cognee MCP server" readme = "README.md" requires-python = ">=3.10" dependencies = [ # For local cognee repo usage remove comment bellow and add absolute path to cognee. Then run `uv sync --reinstall` in the mcp folder on local cognee changes. #"cognee[postgres,codegraph,gemini,huggingface,docs,neo4j] @ file:/Users/igorilic/Desktop/cognee", "cognee[postgres,codegraph,gemini,huggingface,docs,neo4j]==0.3.4", "fastmcp>=2.10.0,<3.0.0", "mcp>=1.12.0,<2.0.0", "uv>=0.6.3,<1.0.0", ] authors = [ { name = "Boris Arzentar", email = "boris@topoteretes.com" }, { name = "Igor Ilic", email = "igor@topoteretes.com" }, { name = "Laszlo Hajdu", email = "laszlo@topoteretes.com" }, ] [build-system] requires = [ "hatchling", ] build-backend = "hatchling.build" [tool.hatch.build.targets.wheel] packages = ["src"] [dependency-groups] dev = [ "debugpy>=1.8.12,<2.0.0", ] [tool.hatch.metadata] allow-direct-references = true [project.scripts] cognee = "src:main" cognee-mcp = "src:main_mcp"

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/topoteretes/cognee'

If you have feedback or need assistance with the MCP directory API, please join our Discord server