Multi-Model Advisor

by YuChenSSR
Verified
# Multi-Model Advisor ## (锵锵四人行) [![smithery badge](https://smithery.ai/badge/@YuChenSSR/multi-ai-advisor-mcp)](https://smithery.ai/server/@YuChenSSR/multi-ai-advisor-mcp) A Model Context Protocol (MCP) server that queries multiple Ollama models and combines their responses, providing diverse AI perspectives on a single question. This creates a "council of advisors" approach where Claude can synthesize multiple viewpoints alongside its own to provide more comprehensive answers. <a href="https://glama.ai/mcp/servers/@YuChenSSR/multi-ai-advisor-mcp"> <img width="380" height="200" src="https://glama.ai/mcp/servers/@YuChenSSR/multi-ai-advisor-mcp/badge" alt="Multi-Model Advisor MCP server" /> </a> ```mermaid graph TD A[Start] --> B[Worker Local AI 1 Opinion] A --> C[Worker Local AI 2 Opinion] A --> D[Worker Local AI 3 Opinion] B --> E[Manager AI] C --> E D --> E E --> F[Decision Made] ``` ## Features - Query multiple Ollama models with a single question - Assign different roles/personas to each model - View all available Ollama models on your system - Customize system prompts for each model - Configure via environment variables - Integrate seamlessly with Claude for Desktop ## Prerequisites - Node.js 16.x or higher - Ollama installed and running (see [Ollama installation](https://github.com/ollama/ollama#installation)) - Claude for Desktop (for the complete advisory experience) ## Installation ### Installing via Smithery To install multi-ai-advisor-mcp for Claude Desktop automatically via [Smithery](https://smithery.ai/server/@YuChenSSR/multi-ai-advisor-mcp): ```bash npx -y @smithery/cli install @YuChenSSR/multi-ai-advisor-mcp --client claude ``` ### Manual Installation 1. Clone this repository: ```bash git clone https://github.com/yourusername/multi-model-advisor.git cd multi-model-advisor ``` 2. Install dependencies: ```bash npm install ``` 3. Build the project: ```bash npm run build ``` 4. Install required Ollama models: ```bash ollama pull gemma3:1b ollama pull llama3.2:1b ollama pull deepseek-r1:1.5b ``` ## Configuration Create a `.env` file in the project root with your desired configuration: ``` # Server configuration SERVER_NAME=multi-model-advisor SERVER_VERSION=1.0.0 DEBUG=true # Ollama configuration OLLAMA_API_URL=http://localhost:11434 DEFAULT_MODELS=gemma3:1b,llama3.2:1b,deepseek-r1:1.5b # System prompts for each model GEMMA_SYSTEM_PROMPT=You are a creative and innovative AI assistant. Think outside the box and offer novel perspectives. LLAMA_SYSTEM_PROMPT=You are a supportive and empathetic AI assistant focused on human well-being. Provide considerate and balanced advice. DEEPSEEK_SYSTEM_PROMPT=You are a logical and analytical AI assistant. Think step-by-step and explain your reasoning clearly. ``` ## Connect to Claude for Desktop 1. Locate your Claude for Desktop configuration file: - MacOS: `~/Library/Application Support/Claude/claude_desktop_config.json` - Windows: `%APPDATA%\Claude\claude_desktop_config.json` 2. Edit the file to add the Multi-Model Advisor MCP server: ```json { "mcpServers": { "multi-model-advisor": { "command": "node", "args": ["/absolute/path/to/multi-model-advisor/build/index.js"] } } } ``` 3. Replace `/absolute/path/to/` with the actual path to your project directory 4. Restart Claude for Desktop ## Usage Once connected to Claude for Desktop, you can use the Multi-Model Advisor in several ways: ### List Available Models You can see all available models on your system: ``` Show me which Ollama models are available on my system ``` This will display all installed Ollama models and indicate which ones are configured as defaults. ### Basic Usage Simply ask Claude to use the multi-model advisor: ``` what are the most important skills for success in today's job market, you can use gemma3:1b, llama3.2:1b, deepseek-r1:1.5b to help you ``` Claude will query all default models and provide a synthesized response based on their different perspectives. ![example](https://raw.githubusercontent.com/YuChenSSR/pics/master/imgs/2025-03-24/Q53YEwdTaeTuL6a7.png) ## How It Works 1. The MCP server exposes two tools: - `list-available-models`: Shows all Ollama models on your system - `query-models`: Queries multiple models with a question 2. When you ask Claude a question referring to the multi-model advisor: - Claude decides to use the `query-models` tool - The server sends your question to multiple Ollama models - Each model responds with its perspective - Claude receives all responses and synthesizes a comprehensive answer 3. Each model can have a different "persona" or role assigned, encouraging diverse perspectives. ## Troubleshooting ### Ollama Connection Issues If the server can't connect to Ollama: - Ensure Ollama is running (`ollama serve`) - Check that the OLLAMA_API_URL is correct in your .env file - Try accessing http://localhost:11434 in your browser to verify Ollama is responding ### Model Not Found If a model is reported as unavailable: - Check that you've pulled the model using `ollama pull <model-name>` - Verify the exact model name using `ollama list` - Use the `list-available-models` tool to see all available models ### Claude Not Showing MCP Tools If the tools don't appear in Claude: - Ensure you've restarted Claude after updating the configuration - Check the absolute path in claude_desktop_config.json is correct - Look at Claude's logs for error messages ### RAM is not enough Some managers' AI models may have chosen larger models, but there is not enough memory to run them. You can try specifying a smaller model (see the [Basic Usage](#basic-usage)) or upgrading the memory. ## License MIT License For more details, please see the LICENSE file in [this project repository](https://github.com/YuChenSSR/multi-ai-advisor-mcp) ## Contributing Contributions are welcome! Please feel free to submit a Pull Request.