Skip to main content
Glama

Gemini MCP Server

by D-Diaa
README.md1.56 kB
# Gemini MCP Server A Model Context Protocol (MCP) server that enables Claude to interact with Google's Gemini AI models. Inspired from [this GitHub repo](https://github.com/RaiAnsar/claude_code-gemini-mcp/tree/main) ## Setup 1. **Get a Gemini API key** from [Google AI Studio](https://aistudio.google.com/apikey) 2. **Add to Claude Desktop config** (`~/Library/Application Support/Claude/claude_desktop_config.json`): ```json { "mcpServers": { "gemini": { "command": "uvx", "args": ["--from", "/path/to/gemini-mcp", "gemini-mcp"], "env": { "GEMINI_API_KEY": "your_api_key_here", "MODEL_NAME": "gemini-2.5-flash-preview-05-20" } } } } ``` Replace `/path/to/gemini-mcp` with the actual path to this directory. ## Usage ### `ask_gemini` Ask Gemini questions directly from Claude. - `prompt`: Your question for Gemini - `temperature`: Response creativity (0.0-1.0, default: 0.5) - `context`: Additional context - `persona`: Role for Gemini (e.g., "senior architect", "security expert") ### `server_info` Check server status and Gemini availability. ## Examples ``` Ask Gemini: What are the latest trends in machine learning? Ask Gemini to review this code as a senior developer: [code] Ask Gemini about Python best practices with context that I'm building a web API ``` ## Troubleshooting - **API key error**: Set `GEMINI_API_KEY` in your environment - **Rate limit**: Wait and try again - **Content filtered**: Rephrase your request - **Server issues**: Use `server_info` tool to check status

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/D-Diaa/gemini-mcp'

If you have feedback or need assistance with the MCP directory API, please join our Discord server