Skip to main content
Glama

Gemini Docs MCP Server

An local STDIO MCP server that provides tools to search and retrieve Google Gemini API documentation.

  • Search Documentation: Full-text search across all Gemini documentation pages.

  • Get Capabilities: List available documentation pages or retrieve content for a specific page.

  • Get Current Model: Quickly access documentation for current Gemini models.

  • Automatic Updates: Scrapes and updates documentation on server startup.

sequenceDiagram participant Client as MCP Client / IDE participant Server as FastMCP Server participant DB as SQLite Database Client->>Server: call_tool("search_documentation", queries=["embeddings"]) Server->>DB: Full-Text Search for "embeddings" DB-->>Server: Return matching documentation Server-->>Client: Return formatted results

How it Works

  1. Ingestion: On startup, the server fetches https://ai.google.dev/gemini-api/docs/llms.txt to get a list of all available documentation pages.

  2. Processing: It then concurrently fetches and processes each page, extracting the text content.

  3. Indexing: The processed content is stored in a local SQLite database with a Full-Text Search (FTS5) index for efficient querying.

  4. Searching: When you use the search_documentation tool, the server queries this SQLite database to find the most relevant documentation pages.

Installation

Option 1: Use uvx (Recommended)

You can use uvx to run the server directly without explicit installation. This is the easiest way to get started.

uvx --from git+https://github.com/philschmid/gemini-api-docs-mcp gemini-docs-mcp

Option 2: Install directly from GitHub

You can install the package directly from GitHub using pip:

pip install git+https://github.com/philschmid/gemini-api-docs-mcp.git

Option 3: Manual Installation (for development)

git clone https://github.com/philschmid/gemini-api-docs-mcp.git cd gemini-api-docs-mcp pip install -e . cd .. rm -rf gemini-api-docs-mcp

Usage

If you installed via pip (Option 2 or 3), run the server using:

gemini-docs-mcp

This will start the MCP server over stdio. It will immediately begin ingesting documentation, which might take a few moments on the first run.

Configuration

The database is stored at ~/.mcp/gemini-api-docs/database.db by default. You can override this by setting the GEMINI_DOCS_DB_PATH environment variable.

Using with an MCP Client

Configure your MCP client to run the gemini-docs-mcp command.

{ "mcpServers": { "gemini-docs": { "command": "uvx", "args": ["--from", "git+https://github.com/philschmid/gemini-api-docs-mcp", "gemini-docs-mcp"] } } }
{ "mcpServers": { "gemini-docs": { "command": "gemini-docs-mcp", } } }

Tools

  • search_documentation(queries: list[str]): Performs a full-text search on Gemini documentation for the given list of queries (max 3).

  • get_capability_page(capability: str = None): Get a list of capabilities or content for a specific one.

  • get_current_model(): Get documentation for current Gemini models.

License

MIT

Test Results

We run a comprehensive evaluation harness to ensure the MCP server provides accurate and up-to-date code examples. The tests cover both Python and TypeScript SDKs.

Metric

Value

Total Tests

117

Passed

114

Failed

3

Last updated: 2025-11-03 13:29:01

You can find the detailed test results in tests/result.json.

Deploy Server
A
security – no known vulnerabilities
A
license - permissive license
A
quality - confirmed to work

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/philschmid/gemini-api-docs-mcp'

If you have feedback or need assistance with the MCP directory API, please join our Discord server