pyproject.toml•971 B
[project]
name = "luma-ai-mcp-server"
version = "0.1.0"
description = "MCP server for Luma AI video generation"
requires-python = ">=3.10"
authors = [
{ name = "Bobby Battista", email = "bobtista@gmail.com" }
]
dependencies = [
"httpx>=0.24.0",
"pydantic>=2.0.0",
"python-dotenv>=1.0.0",
"click>=8.1.0",
"mcp>=0.2.0",
]
[project.optional-dependencies]
test = [
"pytest>=7.0.0",
"pytest-asyncio>=0.21.0",
"pytest-cov>=4.1.0",
]
[build-system]
requires = ["hatchling"]
build-backend = "hatchling.build"
[tool.hatch.build.targets.wheel]
packages = ["src/luma_ai_mcp_server"]
[tool.pytest.ini_options]
asyncio_mode = "auto"
testpaths = ["tests"]
python_files = ["test_*.py"]
addopts = "-v --cov=luma_ai_mcp_server --cov-report=term-missing"
[tool.ruff]
line-length = 100
target-version = "py39"
[tool.ruff.lint]
select = ["E", "F", "I", "N", "W", "B", "UP"]
[project.scripts]
luma-ai-mcp-server = "luma_ai_mcp_server.__main__:main"