We provide all the information about MCP servers via our MCP API.
curl -X GET 'https://glama.ai/api/mcp/v1/servers/Vijay-2005/gemini_mcp'
If you have feedback or need assistance with the MCP directory API, please join our Discord server
# Smithery configuration file: https://smithery.ai/docs/config#smitheryyaml
runtime: "container"
build:
dockerfile: "Dockerfile"
dockerBuildPath: "."
startCommand:
type: "http"
configSchema:
type: "object"
properties:
geminiApiKey:
type: "string"
description: "Your Google AI Studio API key"
defaultModel:
type: "string"
default: "gemini-pro"
description: "Default Gemini model to use"
defaultTemperature:
type: "number"
default: 0.7
description: "Temperature value for generation"
maxOutputTokens:
type: "number"
default: 1000
description: "Maximum number of tokens to generate"
required: ["geminiApiKey"]
exampleConfig:
geminiApiKey: "your-gemini-api-key-here"
defaultModel: "gemini-pro"
defaultTemperature: 0.7
maxOutputTokens: 1000