Multi LLM Cross-Check MCP Server
A Model Control Protocol (MCP) server that allows cross-checking responses from multiple LLM providers simultaneously. This server integrates with Claude Desktop as an MCP server to provide a unified interface for querying different LLM APIs.
Features
- Query multiple LLM providers in parallel
- Currently supports:
- OpenAI (ChatGPT)
- Anthropic (Claude)
- Perplexity AI
- Google (Gemini)
- Asynchronous parallel processing for faster responses
- Easy integration with Claude Desktop
Prerequisites
- Python 3.8 or higher
- API keys for the LLM providers you want to use
- uv package manager (install with
pip install uv
)
Installation
- Clone this repository:
- Initialize uv environment and install requirements:
- Configure in Claude Desktop:
Create a file named
claude_desktop_config.json
in your Claude Desktop configuration directory with the following content:Notes:Copy- You only need to add the API keys for the LLM providers you want to use. The server will skip any providers without configured API keys.
- You may need to put the full path to the uv executable in the command field. You can get this by running
which uv
on MacOS/Linux orwhere uv
on Windows.
Using the MCP Server
Once configured:
- The server will automatically start when you open Claude Desktop
- You can use the
cross_check
tool in your conversations by asking to "cross check with other LLMs" - Provide a prompt, and it will return responses from all configured LLM providers
API Response Format
The server returns a dictionary with responses from each LLM provider:
Error Handling
- If an API key is not provided for a specific LLM, that provider will be skipped
- API errors are caught and returned in the response
- Each LLM's response is independent, so errors with one provider won't affect others
Contributing
Contributions are welcome! Please feel free to submit a Pull Request.
License
This project is licensed under the MIT License - see the LICENSE file for details.
This server cannot be installed
remote-capable server
The server can be hosted and run remotely because it primarily relies on remote services or has no dependency on the local environment.
A Model Control Protocol server that integrates with Claude Desktop to enable simultaneous querying and cross-checking of responses from multiple LLM providers including OpenAI, Anthropic, Perplexity AI, and Google Gemini.
Related MCP Servers
- -securityAlicense-qualityA Model Context Protocol (MCP) server that lets you seamlessly use OpenAI's models right from Claude.Last updated -12428JavaScriptMIT License
- AsecurityAlicenseAqualityA Model Context Protocol server that provides LLMs with access to Valyu's knowledge retrieval and feedback APIs for searching proprietary/web sources and submitting transaction feedback.Last updated -2TypeScriptMIT License
- -securityAlicense-qualityA Model Context Protocol server that enables LLMs like Claude to interact with SQLite and SQL Server databases, allowing for schema inspection and SQL query execution.Last updated -19810TypeScriptMIT License
- -securityFlicense-qualityA unified Model Context Protocol Gateway that bridges LLM interfaces with various tools and services, providing OpenAI API compatibility and supporting both synchronous and asynchronous tool execution.Last updated -Python