Skip to main content
Glama

Gemini MCP Server

listmodels.mdโ€ข3.96 kB
# ListModels Tool - List Available Models **Display all available AI models organized by provider** The `listmodels` tool shows which providers are configured, available models, their aliases, context windows, and capabilities. This is useful for understanding what models can be used and their characteristics. ## Usage ``` "Use zen to list available models" ``` ## Key Features - **Provider organization**: Shows all configured providers and their status - **Model capabilities**: Context windows, thinking mode support, and special features - **Alias mapping**: Shows shorthand names and their full model mappings - **Configuration status**: Indicates which providers are available based on API keys - **Context window information**: Helps you choose models based on your content size needs - **Capability overview**: Understanding which models support extended thinking, vision, etc. ## Output Information The tool displays: **Provider Status:** - Which providers are configured and available - API key status (without revealing the actual keys) - Provider priority order **Model Details:** - Full model names and their aliases - Context window sizes (tokens) - Special capabilities (thinking modes, vision support, etc.) - Provider-specific features **Capability Summary:** - Which models support extended thinking - Vision-capable models for image analysis - Models with largest context windows - Fastest models for quick tasks ## Example Output ``` ๐Ÿ“‹ Available Models by Provider ๐Ÿ”น Google (Gemini) - โœ… Configured โ€ข pro (gemini-2.5-pro) - 1M context, thinking modes โ€ข flash (gemini-2.0-flash-experimental) - 1M context, ultra-fast ๐Ÿ”น OpenAI - โœ… Configured โ€ข o3 (o3) - 200K context, strong reasoning โ€ข o3-mini (o3-mini) - 200K context, balanced โ€ข o4-mini (o4-mini) - 200K context, latest reasoning ๐Ÿ”น Custom/Local - โœ… Configured โ€ข local-llama (llama3.2) - 128K context, local inference โ€ข Available at: http://localhost:11434/v1 ๐Ÿ”น OpenRouter - โŒ Not configured Set OPENROUTER_API_KEY to enable access to Claude, GPT-4, and more models ``` ## When to Use ListModels - **Model selection**: When you're unsure which models are available - **Capability checking**: To verify what features each model supports - **Configuration validation**: To confirm your API keys are working - **Context planning**: To choose models based on content size requirements - **Performance optimization**: To select the right model for speed vs quality trade-offs ## Configuration Dependencies The available models depend on your configuration: **API Keys Required:** - `GEMINI_API_KEY` - Enables Gemini Pro and Flash models - `OPENAI_API_KEY` - Enables OpenAI O3, O4-mini, and GPT models - `OPENROUTER_API_KEY` - Enables access to multiple providers through OpenRouter - `CUSTOM_API_URL` - Enables local/custom models (Ollama, vLLM, etc.) **Model Restrictions:** If you've set model usage restrictions via environment variables, the tool will show: - Which models are allowed vs restricted - Active restriction policies - How to modify restrictions ## Tool Parameters This tool requires no parameters - it simply queries the server configuration and displays all available information. ## Best Practices - **Check before planning**: Use this tool to understand your options before starting complex tasks - **Verify configuration**: Confirm your API keys are working as expected - **Choose appropriate models**: Match model capabilities to your specific needs - **Understand limits**: Be aware of context windows when working with large files ## When to Use ListModels vs Other Tools - **Use `listmodels`** for: Understanding available options and model capabilities - **Use `chat`** for: General discussions about which model to use for specific tasks - **Use `version`** for: Server configuration and version information - **Use other tools** for: Actual analysis, debugging, or development work

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/BeehiveInnovations/gemini-mcp-server'

If you have feedback or need assistance with the MCP directory API, please join our Discord server