Skip to main content
Glama

Gemini MCP Server

by lbds137
pyproject.toml1.51 kB
[build-system] requires = ["setuptools>=45", "wheel"] build-backend = "setuptools.build_meta" [tool.black] line-length = 100 target-version = ['py39', 'py310', 'py311', 'py312'] include = '\.pyi?$' extend-exclude = ''' /( # directories \.eggs | \.git | \.hg | \.mypy_cache | \.tox | \.venv | build | dist )/ ''' [tool.isort] profile = "black" line_length = 100 multi_line_output = 3 include_trailing_comma = true force_grid_wrap = 0 use_parentheses = true ensure_newline_before_comments = true [tool.mypy] python_version = "3.9" warn_return_any = false warn_unused_configs = true disallow_untyped_defs = false disallow_incomplete_defs = false check_untyped_defs = true disallow_untyped_decorators = false no_implicit_optional = true warn_redundant_casts = true warn_unused_ignores = true warn_no_return = true warn_unreachable = false strict_equality = true [[tool.mypy.overrides]] module = "google.*" ignore_missing_imports = true [[tool.mypy.overrides]] module = "pytest" ignore_missing_imports = true [tool.pytest.ini_options] minversion = "6.0" testpaths = ["tests"] python_files = "test_*.py" python_classes = "Test*" python_functions = "test_*" addopts = "-v --tb=short --strict-markers" [tool.coverage.run] source = ["src/gemini_mcp"] omit = ["*/tests/*", "*/__init__.py"] [tool.coverage.report] exclude_lines = [ "pragma: no cover", "def __repr__", "raise AssertionError", "raise NotImplementedError", "if __name__ == .__main__.:", "if TYPE_CHECKING:", ]

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/lbds137/gemini-mcp-server'

If you have feedback or need assistance with the MCP directory API, please join our Discord server