Provides integration with Google's Gemini AI models, enabling access to models like gemini-2.5-flash through Google's generative language API.
Gemimi MCP Server (in Python)
Model Context Protocol (MCP) server for Gemimi integration, built on FastMCP.
This server is implemented in Python, with fastmcp.
Quick Start
- Build the Docker image:
Integration with Cursor/Claude
In MCP Settings -> Add MCP server, add this config:
Note: Don't forget to replace GEMINI_API_KEY
、GEMINI_MODEL
、GEMINI_BASE_URL
、HTTP_PROXY
、HTTPS_PROXY
values with your actual Gemimi credentials and instance URL.
This server cannot be installed
remote-capable server
The server can be hosted and run remotely because it primarily relies on remote services or has no dependency on the local environment.
A Python-based MCP server that enables integration of Gemini AI models with MCP-compatible applications like Cursor/Claude, allowing for interaction with Gemini APIs through the Model Context Protocol.
Related MCP Servers
- -securityAlicense-qualityModel Context Protocol (MCP) server implementation that enables Claude Desktop to interact with Google's Gemini AI models.Last updated -182JavaScriptMIT License
- -security-license-qualityAn MCP server implementation that allows using Google's Gemini AI models (specifically Gemini 1.5 Pro) through Claude or other MCP clients via the Model Context Protocol.Last updated -1JavaScript
- AsecurityAlicenseAqualityA dedicated server that wraps Google's Gemini AI models in a Model Context Protocol (MCP) interface, allowing other LLMs and MCP-compatible systems to access Gemini's capabilities like content generation, function calling, chat, and file handling through standardized tools.Last updated -1629TypeScriptMIT License
- -securityAlicense-qualityAn MCP server that enables other AI models (like Claude) to use Google's Gemini models as tools for specific tasks through a standardized interface.Last updated -1TypeScriptMIT License