Skip to main content
Glama

MCP Gemini Server

by Vijay-2005

MseeP.ai Security Assessment Badge

MCP Gemini Server

This MCP server allows you to access Google's Gemini API directly from Claude Desktop.

Features

  • Call the Gemini API with customizable parameters

  • Ask Claude and Gemini to talk to each other in a long running discussion!

  • Configure model versions, temperature, and other parameters

  • Access to various Gemini models (gemini-pro, gemini-1.5-pro, gemini-1.5-flash, etc.)

  • Conversation history management for context-aware responses

  • Use your own Google AI Studio API key

Setup Instructions

Installing via Smithery

To install Gemini Server for Claude Desktop via Smithery:

# Install dependencies npm install # Deploy to Smithery npm run deploy

After deployment, you should be able to access your MCP server at:

https://smithery.ai/server/@YOUR_USERNAME/gemini_mcp/mcp/manifest

Prerequisites

Installation for Local Development

  1. Clone this repository:

    git clone https://github.com/your-username/gemini_mcp.git cd gemini_mcp
  2. Install dependencies:

    npm install
  3. Set your Gemini API key in the Smithery dashboard or through the config file.

  4. Deploy to Smithery:

    npm run deploy

Using with Claude Desktop

  1. Configure Claude Desktop to use this MCP server by following the instructions at: MCP Quickstart Guide

  2. Add the following configuration to your Claude Desktop settings:

  3. Restart Claude Desktop.

  4. You can now use the Gemini API through Claude by asking questions that mention Gemini.

Available Tools

The MCP server provides the following tools:

  1. ask_gemini(prompt, model, temperature, max_output_tokens, conversation_id) - Send a prompt to Gemini and get a response

  2. ask_gemini_with_web_search(prompt, model, temperature, max_output_tokens, conversation_id) - Send a prompt to Gemini with enhanced search instructions

Example Usage

Basic Gemini usage:

Tell Claude to ask Gemini a question!

Use the ask_gemini tool to answer: What is the best way to learn Python?

Tell Claude to have a conversation with Gemini:

Use the ask_gemini tool to have a two way conversation between you and Gemini about the topic that is most important to you.

Note how the conversation_id allows maintaining conversation history for context-aware responses across multiple interactions.

With search capability:

For questions that may benefit from comprehensive information:

Use the ask_gemini_with_web_search tool to answer: What are the latest developments in quantum computing?

Try search-enhanced responses for planning:

Use the ask_gemini_with_web_search tool to find information about weather patterns and based on that, keep using the tool to build up a great day out for someone who loves food and parks

How It Works

This tool utilizes Google's Gemini API with conversation state management. This approach:

  1. Maintains conversation history locally for context tracking

  2. Provides access to various Gemini models (pro, flash, vision, etc.)

  3. Improves the user experience by maintaining context across messages

  4. Allows enhanced responses with search-oriented prompting

License

MIT License

-
security - not tested
F
license - not found
-
quality - not tested

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/Vijay-2005/gemini_mcp'

If you have feedback or need assistance with the MCP directory API, please join our Discord server