Multi LLM Cross-Check MCP Server
A Model Control Protocol (MCP) server that allows cross-checking responses from multiple LLM providers simultaneously. This server integrates with Claude Desktop as an MCP server to provide a unified interface for querying different LLM APIs.
Features
- Query multiple LLM providers in parallel
- Currently supports:
- OpenAI (ChatGPT)
- Anthropic (Claude)
- Perplexity AI
- Google (Gemini)
- Asynchronous parallel processing for faster responses
- Easy integration with Claude Desktop
Prerequisites
- Python 3.8 or higher
- API keys for the LLM providers you want to use
- uv package manager (install with
pip install uv
)
Installation
- Clone this repository:
Copy
- Initialize uv environment and install requirements:
Copy
- Configure in Claude Desktop:
Create a file named
claude_desktop_config.json
in your Claude Desktop configuration directory with the following content:Notes:Copy- You only need to add the API keys for the LLM providers you want to use. The server will skip any providers without configured API keys.
- You may need to put the full path to the uv executable in the command field. You can get this by running
which uv
on MacOS/Linux orwhere uv
on Windows.
Using the MCP Server
Once configured:
- The server will automatically start when you open Claude Desktop
- You can use the
cross_check
tool in your conversations by asking to "cross check with other LLMs" - Provide a prompt, and it will return responses from all configured LLM providers
API Response Format
The server returns a dictionary with responses from each LLM provider:
Copy
Error Handling
- If an API key is not provided for a specific LLM, that provider will be skipped
- API errors are caught and returned in the response
- Each LLM's response is independent, so errors with one provider won't affect others
Contributing
Contributions are welcome! Please feel free to submit a Pull Request.
License
This project is licensed under the MIT License - see the LICENSE file for details.
This server cannot be installed
A Model Control Protocol server that integrates with Claude Desktop to enable simultaneous querying and cross-checking of responses from multiple LLM providers including OpenAI, Anthropic, Perplexity AI, and Google Gemini.