Skip to main content
Glama

Gemini MCP Tool

models.md1.58 kB
# Model Selection Choose the right Gemini model for your task. ## Available Models ### Gemini-2.5-pro - **Best for**: Complex analysis, large codebases - **Context**: 2M tokens - **Use when**: Analyzing entire projects, architectural reviews, stronger reasoning ### Gemini-2.5-flash - **Best for**: Quick responses, routine tasks - **Context**: 1M tokens - **Use when**: Fast code reviews, Analyzing entire projects, simple explanations ## Setting Models ```bash You need use natural language: "...using gemini flash" ``` ```bash You can also append with '-m' or ask specifically with ``` ### In Configuration ```json { "mcpServers": { "gemini-cli": { "command": "gemini-mcp", "env": { "GEMINI_MODEL": "gemini-1.5-flash" } } } } ``` ### Per Request (Coming Soon) ``` /gemini-cli:analyze --model=flash @file.js quick review ``` ## Model Comparison | Model | Speed | Context | Best Use Case | |-------|-------|---------|---------------| | Pro | Slower | 2M tokens | big ideas | | Flash | Fast | 1M tokens | quick, specific changes | ## Cost Optimization 1. **Start with Flash** for most tasks 2. **Use Pro** only when you need the full context 3. **Flash-8B** for simple, repetitive tasks ## Token Limits - **Pro**: ~2 million tokens (~500k lines of code) - **Flash**: ~1 million tokens (~250k lines of code) - **Flash-8B**: ~1 million tokens (~250k lines of code) ## Recommendations - **Code Review**: Flash - **Architecture Analysis**: Pro - **Quick Fixes**: Flash-8B - **Documentation**: Flash - **Security Audit**: Pro

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/jamubc/gemini-mcp-tool'

If you have feedback or need assistance with the MCP directory API, please join our Discord server