Skip to main content
Glama

Multi LLM Cross-Check MCP Server

smithery.yaml1.17 kB
# Smithery configuration file: https://smithery.ai/docs/build/project-config startCommand: type: stdio commandFunction: # A JS function that produces the CLI command based on the given config to start the MCP on stdio. |- (config) => ({ command: 'python', args: ['main.py'], env: { OPENAI_API_KEY: config.openaiApiKey || '', ANTHROPIC_API_KEY: config.anthropicApiKey || '', PERPLEXITY_API_KEY: config.perplexityApiKey || '', GEMINI_API_KEY: config.geminiApiKey || '' } }) configSchema: # JSON Schema defining the configuration options for the MCP. type: object properties: openaiApiKey: type: string description: API key for OpenAI (ChatGPT) anthropicApiKey: type: string description: API key for Anthropic (Claude) perplexityApiKey: type: string description: API key for Perplexity AI geminiApiKey: type: string description: API key for Google Gemini exampleConfig: openaiApiKey: sk-xxxxxxxxxxxxxxxxxxxxxxxxxxxx anthropicApiKey: anthropic-xxxxxxxxxxxxxxxxxxxx perplexityApiKey: perplexity-xxxxxxxxxxxxxxxxxx geminiApiKey: AIzaSy...

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/lior-ps/multi-llm-cross-check-mcp-server'

If you have feedback or need assistance with the MCP directory API, please join our Discord server