Claude Concilium
Multi-agent AI consultation framework for Claude Code via MCP.
Get a second (and third) opinion from other LLMs when Claude Code alone isn't enough.
The Problem
Claude Code is powerful, but one brain can miss bugs, overlook edge cases, or get stuck in a local optimum. Critical decisions benefit from diverse perspectives.
The Solution
Concilium runs parallel consultations with multiple LLMs through standard MCP protocol. Each LLM server wraps a CLI tool — no API keys needed for the primary providers (they use OAuth).
Key features:
Parallel consultation with 2+ AI agents
Production-grade fallback chains with error detection
Each MCP server works standalone or as part of Concilium
Plug & play: clone,
npm install, add to.mcp.json
Architecture
Quickstart
1. Clone and install
Expected output:
2. Set up providers
Pick at least 2 providers:
Provider | Auth | Free Tier | Setup |
OpenAI |
| ChatGPT Plus weekly credits | |
Gemini | Google OAuth | 1000 req/day | |
Qwen |
| Varies | |
DeepSeek | API key | Pay-per-use (cheap) |
3. Add to Claude Code
Copy config/mcp.json.example and update paths:
Or add servers individually to your existing .mcp.json:
4. Install the skill (optional)
Copy the Concilium skill to your Claude Code commands:
Now use /ai-concilium in Claude Code to trigger a multi-agent consultation.
MCP Servers
Each server can be used independently — you don't need all of them.
Server | CLI Tool | Auth | Tools |
| OAuth (ChatGPT Plus) |
| |
| Google OAuth |
| |
| API key / CLI login |
|
DeepSeek uses the existing deepseek-mcp-server npm package — no custom server needed.
How It Works
Consultation Flow
Formulate — describe the problem concisely (under 500 chars)
Send in parallel — OpenAI + Gemini get the same prompt
Handle errors — if a provider fails, fallback chain kicks in (Qwen → DeepSeek)
Synthesize — compare responses, find consensus
Iterate (optional) — resolve disagreements with follow-up questions
Decide — apply the synthesized solution
Error Detection
All servers detect provider-specific errors and return structured responses:
Error Type | Meaning | Action |
| Rate/credit limit hit | Use fallback provider |
| Token needs refresh | Re-authenticate CLI |
| Model unavailable on plan | Use default model |
Timeout | Process hung | Auto-killed, use fallback |
Fallback Chain
When to Use Concilium
Scenario | Recommended Agents |
Code review | OpenAI + Gemini (parallel) |
Architecture decision | OpenAI + Gemini → iterate if disagree |
Stuck bug (3+ attempts) | All available agents |
Performance optimization | Gemini (1M context) + OpenAI |
Security review | OpenAI + Gemini + manual verification |
Customization
See docs/customization.md for:
Adding your own LLM provider
Modifying the fallback chain
MCP server template
Custom prompt strategies
Documentation
Architecture — flow diagrams, error handling, design decisions
OpenAI Setup — Codex CLI, ChatGPT Plus, minimal config
Gemini Setup — gemini-cli, Google OAuth
Qwen Setup — Qwen CLI, DashScope
DeepSeek Setup — API key, npm package
Customization — add your own LLM, modify chains
License
MIT