Provides integration with Google's Gemini AI models, enabling access to models like gemini-2.5-flash through Google's generative language API.
Gemimi MCP Server (in Python)
Model Context Protocol (MCP) server for Gemimi integration, built on FastMCP.
This server is implemented in Python, with fastmcp.
Quick Start
- Build the Docker image:
Integration with Cursor/Claude
In MCP Settings -> Add MCP server, add this config:
Note: Don't forget to replace GEMINI_API_KEY
、GEMINI_MODEL
、GEMINI_BASE_URL
、HTTP_PROXY
、HTTPS_PROXY
values with your actual Gemimi credentials and instance URL.
This server cannot be installed
A Python-based MCP server that enables integration of Gemini AI models with MCP-compatible applications like Cursor/Claude, allowing for interaction with Gemini APIs through the Model Context Protocol.
Related MCP Servers
- -securityAlicense-qualityModel Context Protocol (MCP) server implementation that enables Claude Desktop to interact with Google's Gemini AI models.Last updated -53TypeScriptMIT License
- -security-license-qualityAn MCP server implementation that allows using Google's Gemini AI models (specifically Gemini 1.5 Pro) through Claude or other MCP clients via the Model Context Protocol.Last updated -1JavaScript
- AsecurityAlicenseAqualityA dedicated server that wraps Google's Gemini AI models in a Model Context Protocol (MCP) interface, allowing other LLMs and MCP-compatible systems to access Gemini's capabilities like content generation, function calling, chat, and file handling through standardized tools.Last updated -1620TypeScriptMIT License
- -securityAlicense-qualityAn MCP server that enables other AI models (like Claude) to use Google's Gemini models as tools for specific tasks through a standardized interface.Last updated -1TypeScriptMIT License