Supports environment variables through .env file for storing sensitive information like API keys securely.
Provides repository access through GitHub for cloning and installing the server.
Enables interaction with Google's Gemini API, allowing text generation, text analysis, and chat conversations through the Gemini models.
Supports running on Linux with specific virtual environment activation instructions.
Supports running on macOS with specific virtual environment activation instructions.
Supports testing and debugging API requests using Postman for manual testing.
Built using Python, requiring version 3.7 or higher for server implementation.
Click on "Install Server".
Wait a few minutes for the server to deploy. Once ready, it will show a "Started" state.
In the chat, type
@followed by the MCP server name and your instructions, e.g., "@MCP Gemini Serveranalyze the sentiment of this customer review: 'The product arrived damaged and customer service was unhelpful.'"
That's it! The server will respond to your query, and you can continue using it as needed.
Here is a step-by-step guide with screenshots.
MCP Gemini Server
A server implementation of the Model Context Protocol (MCP) to enable AI assistants like Claude to interact with Google's Gemini API.
Project Overview
This project implements a server that follows the Model Context Protocol, allowing AI assistants to communicate with Google's Gemini models. With this MCP server, AI assistants can request text generation, text analysis, and maintain chat conversations through the Gemini API.
Related MCP server: Gemini MCP Server
Features
Client-Server Communication: Implements MCP protocol for secure message exchange between client and server.
Message Processing: Handles and processes client requests, sending appropriate responses.
Error Handling & Logging: Logs server activities and ensures smooth error recovery.
Environment Variables Support: Uses
.envfile for storing sensitive information securely.API Testing & Debugging: Supports manual and automated testing using Postman and test scripts.
Installation
Prerequisites
Python 3.7 or higher
Google AI API key
Setup
Clone this repository:
git clone https://github.com/yourusername/mcp-gemini-server.git
cd mcp-gemini-serverCreate a virtual environment:
python -m venv venvActivate the virtual environment:
Windows:
venv\Scripts\activatemacOS/Linux:
source venv/bin/activate
Install dependencies:
pip install -r requirements.txtCreate a
.envfile in the root directory with your Gemini API key:
GEMINI_API_KEY=your_api_key_hereUsage
Start the server:
python server.pyThe server will run on
http://localhost:5000/by defaultSend MCP requests to the
/mcpendpoint using POST method
Example Request
import requests
url = 'http://localhost:5000/mcp'
payload = {
'action': 'generate_text',
'parameters': {
'prompt': 'Write a short poem about AI',
'temperature': 0.7
}
}
response = requests.post(url, json=payload)
print(response.json())API Reference
Endpoints
GET /health: Check if the server is runningGET /list-models: List available Gemini modelsPOST /mcp: Main endpoint for MCP requests
MCP Actions
1. generate_text
Generate text content with Gemini.
Parameters:
prompt(required): The text prompt for generationtemperature(optional): Controls randomness (0.0 to 1.0)max_tokens(optional): Maximum tokens to generate
Example:
{
"action": "generate_text",
"parameters": {
"prompt": "Write a short story about a robot",
"temperature": 0.8,
"max_tokens": 500
}
}2. analyze_text
Analyze text content.
Parameters:
text(required): The text to analyzeanalysis_type(optional): Type of analysis ('sentiment', 'summary', 'keywords', or 'general')
Example:
{
"action": "analyze_text",
"parameters": {
"text": "The weather today is wonderful! I love how the sun is shining.",
"analysis_type": "sentiment"
}
}3. chat
Have a conversation with Gemini.
Parameters:
messages(required): Array of message objects with 'role' and 'content'temperature(optional): Controls randomness (0.0 to 1.0)
Example:
{
"action": "chat",
"parameters": {
"messages": [
{"role": "user", "content": "Hello, how are you?"},
{"role": "assistant", "content": "I'm doing well! How can I help?"},
{"role": "user", "content": "Tell me about quantum computing"}
],
"temperature": 0.7
}
}Error Handling
The server returns appropriate HTTP status codes and error messages:
200: Successful request400: Bad request (missing or invalid parameters)500: Server error (API issues, etc.)
Testing
Use the included test script to test various functionalities:
# Test all functionalities
python test_client.py
# Test specific functionality
python test_client.py text # Test text generation
python test_client.py analyze # Test text analysis
python test_client.py chat # Test chat functionalityMCP Protocol Specification
The Model Context Protocol implemented here follows these specifications:
Request Format:
action: String specifying the operationparameters: Object containing action-specific parameters
Response Format:
result: Object containing the operation resulterror: String explaining any error (when applicable)
License
MIT License
This server cannot be installed
Resources
Looking for Admin?
Admins can modify the Dockerfile, update the server description, and track usage metrics. If you are the server author, to access the admin panel.