Skip to main content
Glama

Tecton MCP Server

Official
by tecton-ai
pyproject.toml1.67 kB
[project] name = "tecton_mcp" version = "0.2.0" description = "[private preview] Tecton MCP" authors = [{ name = "Tecton, Inc.", email = "support@tecton.ai" }] dependencies = [ "pandas>=2.2.1", "typing_extensions>=4.5", "lancedb==0.18.0", "langchain-core==0.3.47", "langchain==0.3.21", "langchain-community>=0.3.0", "langchain-huggingface>=0.0.3", "sentence-transformers>=2.2.0", "langchain-openai==0.3.9", "openai==1.67.0", "fastmcp>=2.4.0", "debugpy>=1.8.11", "pyarrow==15.0.2", "tecton>=0.8.0a0", "tecton-client", "requests>=2.31.0", "prometheus-client>=0.19.0" ] readme = "README.md" requires-python = ">=3.11,<3.12" license = { text = "MIT" } classifiers = [ "Programming Language :: Python :: 3", "Operating System :: OS Independent", "License :: OSI Approved :: MIT License", ] [project.urls] Homepage = "https://tecton.ai" Documentation = "https://docs.tecton.ai" [project.scripts] tecton_mcp = "tecton_mcp.cli:main" run-integration-tests = "run_integration_tests:main" [tool.hatch.build.targets.wheel] packages = ["src/tecton_mcp"] [tool.hatch.build.targets.wheel.sources] "src" = "" [build-system] requires = ["hatchling"] build-backend = "hatchling.build" [tool.uv] managed = true dev-dependencies = [ "setuptools", "wheel", "pre-commit>=3.7.1", "pylint>=3.2.6", "pytest>=8.0.0", "pytest-lazy-fixture>=0.6.3", "pyyaml>=6.0", ] [tool.uv.workspace] members = ["demo/marshmallow-demo"] [tool.hatch.metadata] allow-direct-references = true [tool.pytest.ini_options] addopts = "-p pytest_cov -vvv --cov=src --cov-report=term-missing:skip-covered"

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/tecton-ai/tecton-mcp'

If you have feedback or need assistance with the MCP directory API, please join our Discord server