Skip to main content
Glama

mcp-luma-dream-machine

pyproject.toml971 B
[project] name = "luma-ai-mcp-server" version = "0.1.0" description = "MCP server for Luma AI video generation" requires-python = ">=3.10" authors = [ { name = "Bobby Battista", email = "bobtista@gmail.com" } ] dependencies = [ "httpx>=0.24.0", "pydantic>=2.0.0", "python-dotenv>=1.0.0", "click>=8.1.0", "mcp>=0.2.0", ] [project.optional-dependencies] test = [ "pytest>=7.0.0", "pytest-asyncio>=0.21.0", "pytest-cov>=4.1.0", ] [build-system] requires = ["hatchling"] build-backend = "hatchling.build" [tool.hatch.build.targets.wheel] packages = ["src/luma_ai_mcp_server"] [tool.pytest.ini_options] asyncio_mode = "auto" testpaths = ["tests"] python_files = ["test_*.py"] addopts = "-v --cov=luma_ai_mcp_server --cov-report=term-missing" [tool.ruff] line-length = 100 target-version = "py39" [tool.ruff.lint] select = ["E", "F", "I", "N", "W", "B", "UP"] [project.scripts] luma-ai-mcp-server = "luma_ai_mcp_server.__main__:main"

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/bobtista/luma-ai-mcp-server'

If you have feedback or need assistance with the MCP directory API, please join our Discord server