Enables pre-commit validation of git changes across multiple repositories, with the ability to detect incomplete changes, security issues, and ensure implementation matches intent.
Integrates with Gemini 2.5 Pro and 2.0 Flash models for extended thinking capabilities, deep analysis, and ultra-fast responses with 1M token context.
Provides access to OpenAI's O3 and O3-mini models for strong logical reasoning and systematic analysis with a 200K token context window.
Uses Redis for AI-to-AI conversation persistence, enabling multi-turn conversations between Claude and other AI models with full context retention.
Zen MCP: Many Workflows. One Context.
AI orchestration for Claude Code - A Model Context Protocol server that gives your CLI of choice (e.g. Claude Code) access to multiple AI models for enhanced code analysis, problem-solving, and collaborative development. Zen works with Claude Code, Gemini CLI, Codex CLI as well as others.
True AI collaboration with conversation continuity - Claude stays in control but gets perspectives from the best AI for each subtask. Context carries forward seamlessly across tools and models, enabling complex workflows like: code reviews with multiple models → automated planning → implementation → pre-commit validation.
You're in control. Claude orchestrates the AI team, but you decide the workflow. Craft powerful prompts that bring in Gemini Pro, GPT 5, Flash, or local offline models exactly when needed.
Multi-Model Orchestration - Claude coordinates with Gemini Pro, O3, GPT-5, and 50+ other models to get the best analysis for each task
Context Revival Magic - Even after Claude's context resets, continue conversations seamlessly by having other models "remind" Claude of the discussion
Guided Workflows - Enforces systematic investigation phases that prevent rushed analysis and ensure thorough code examination
Extended Context Windows - Break Claude's limits by delegating to Gemini (1M tokens) or O3 (200K tokens) for massive codebases
True Conversation Continuity - Full context flows across tools and models - Gemini remembers what O3 said 10 steps ago
Model-Specific Strengths - Extended thinking with Gemini Pro, blazing speed with Flash, strong reasoning with O3, privacy with local Ollama
Professional Code Reviews - Multi-pass analysis with severity levels, actionable feedback, and consensus from multiple AI experts
Smart Debugging Assistant - Systematic root cause analysis with hypothesis tracking and confidence levels
Automatic Model Selection - Claude intelligently picks the right model for each subtask (or you can specify)
Vision Capabilities - Analyze screenshots, diagrams, and visual content with vision-enabled models
Local Model Support - Run Llama, Mistral, or other models locally for complete privacy and zero API costs
Bypass MCP Token Limits - Automatically works around MCP's 25K limit for large prompts and responses
The Killer Feature: When Claude's context resets, just ask to "continue with O3" - the other model's response magically revives Claude's understanding without re-ingesting documents!
Example: Multi-Model Code Review Workflow
Perform a codereview using gemini pro and o3 and use planner to generate a detailed plan, implement the fixes and do a final precommit check by continuing from the previous codereview
This triggers a
codereview
workflow where Claude walks the code, looking for all kinds of issuesAfter multiple passes, collects relevant code and makes note of issues along the way
Maintains a
confidence
level betweenexploring
,low
,medium
,high
andcertain
to track how confidently it's been able to find and identify issuesGenerates a detailed list of critical -> low issues
Shares the relevant files, findings, etc with Gemini Pro to perform a deep dive for a second
codereview
Comes back with a response and next does the same with o3, adding to the prompt if a new discovery comes to light
When done, Claude takes in all the feedback and combines a single list of all critical -> low issues, including good patterns in your code. The final list includes new findings or revisions in case Claude misunderstood or missed something crucial and one of the other models pointed this out
It then uses the
planner
workflow to break the work down into simpler steps if a major refactor is requiredClaude then performs the actual work of fixing highlighted issues
When done, Claude returns to Gemini Pro for a
precommit
review
All within a single conversation thread! Gemini Pro in step 11 knows what was recommended by O3 in step 7! Taking that context and review into consideration to aid with its final pre-commit review.
Think of it as Claude Code This MCP isn't magic. It's just super-glue.
Remember: Claude stays in full control — but YOU call the shots. Zen is designed to have Claude engage other models only when needed — and to follow through with meaningful back-and-forth. You're the one who crafts the powerful prompt that makes Claude bring in Gemini, Flash, O3 — or fly solo. You're the guide. The prompter. The puppeteer.
You are the AI - Actually Intelligent.
Recommended AI Stack
For best results, use Claude Code with:
Opus 4.1 - All agentic work and orchestration
Gemini 2.5 Pro - Deep thinking, code reviews, debugging, pre-commit analysis
Quick Start (5 minutes)
Prerequisites: Python 3.10+, Git, uv installed
1. Get API Keys (choose one or more):
OpenRouter - Access multiple models with one API
Gemini - Google's latest models
OpenAI - O3, GPT-5 series
X.AI - Grok models
DIAL - Vendor-agnostic model access
Ollama - Local models (free)
2. Install (choose one):
Option A: Clone and Automatic Setup (recommended)
Option B: Instant Setup with
3. Start Using!
👉 Complete Setup Guide with detailed installation, configuration for Gemini / Codex, and troubleshooting
Core Tools
Note: Each tool comes with its own multi-step workflow, parameters, and descriptions that consume valuable context window space even when not in use. To optimize performance, some tools are disabled by default. See Tool Configuration below to enable them.
Collaboration & Planning (Enabled by default)
chat
- Brainstorm ideas, get second opinions, validate approachesthinkdeep
- Extended reasoning, edge case analysis, alternative perspectivesplanner
- Break down complex projects into structured, actionable plansconsensus
- Get expert opinions from multiple AI models with stance steering
Code Analysis & Quality
debug
- Systematic investigation and root cause analysisprecommit
- Validate changes before committing, prevent regressionscodereview
- Professional reviews with severity levels and actionable feedbackanalyze
(disabled by default - - Understand architecture, patterns, dependencies across entire codebases
Development Tools (Disabled by default -
refactor
- Intelligent code refactoring with decomposition focustestgen
- Comprehensive test generation with edge casessecaudit
- Security audits with OWASP Top 10 analysisdocgen
- Generate documentation with complexity analysis
Utilities
challenge
- Prevent "You're absolutely right!" responses with critical analysistracer
(disabled by default - - Static analysis prompts for call-flow mapping
Default Configuration
To optimize context window usage, only essential tools are enabled by default:
Enabled by default:
chat
,thinkdeep
,planner
,consensus
- Core collaboration toolscodereview
,precommit
,debug
- Essential code quality toolschallenge
- Critical thinking utility
Disabled by default:
analyze
,refactor
,testgen
,secaudit
,docgen
,tracer
Enabling Additional Tools
To enable additional tools, remove them from the DISABLED_TOOLS
list:
Option 1: Edit your .env file
Option 2: Configure in MCP settings
Option 3: Enable all tools
Note:
Essential tools (
version
,listmodels
) cannot be disabledAfter changing tool configuration, restart your Claude session for changes to take effect
Each tool adds to context window usage, so only enable what you need
Key Features
AI Orchestration
Auto model selection - Claude picks the right AI for each task
Multi-model workflows - Chain different models in single conversations
Conversation continuity - Context preserved across tools and models
Context revival - Continue conversations even after context resets
Model Support
Multiple providers - Gemini, OpenAI, X.AI, OpenRouter, DIAL, Ollama
Latest models - GPT-5, Gemini 2.5 Pro, O3, Grok-4, local Llama
Thinking modes - Control reasoning depth vs cost
Vision support - Analyze images, diagrams, screenshots
Developer Experience
Guided workflows - Systematic investigation prevents rushed analysis
Smart file handling - Auto-expand directories, manage token limits
Web search integration - Access current documentation and best practices
Large prompt support - Bypass MCP's 25K token limit
Example Workflows
Multi-model Code Review:
→ Claude reviews code systematically → Consults Gemini Pro → Gets O3's perspective → Creates unified action plan
Collaborative Debugging:
→ Deep investigation → Expert analysis → Solution implementation → Pre-commit validation
Architecture Planning:
→ Structured planning → Multiple expert opinions → Consensus building → Implementation roadmap
👉 Advanced Usage Guide for complex workflows, model configuration, and power-user features
Quick Links
📖 Documentation
Getting Started - Complete setup guide
Tools Reference - All tools with examples
Advanced Usage - Power user features
Configuration - Environment variables, restrictions
🔧 Setup & Support
WSL Setup - Windows users
Troubleshooting - Common issues
Contributing - Code standards, PR process
License
Apache 2.0 License - see LICENSE file for details.
Acknowledgments
Built with the power of Multi-Model AI collaboration 🤝
Actual Intelligence by real Humans
MCP (Model Context Protocol) by Anthropic
Claude Code - Your AI coding orchestrator
Gemini 2.5 Pro & Flash - Extended thinking & fast analysis
OpenAI O3 & GPT-5 - Strong reasoning & latest capabilities
Star History
This server cannot be installed
local-only server
The server can only run on the client's local machine because it depends on local resources.
A Model Context Protocol server that gives Claude access to multiple AI models (Gemini, OpenAI, OpenRouter) for enhanced code analysis, problem-solving, and collaborative development through AI orchestration with conversations that continue across tasks.
Related MCP Servers
- -securityFlicense-qualityA Model Context Protocol server that allows Claude to make API requests on your behalf, providing tools for testing various APIs including HTTP requests and OpenAI integrations without sharing your API keys in the chat.Last updated -
- -securityAlicense-qualityA Model Context Protocol server that enables Claude to collaborate with Google's Gemini AI models, providing tools for question answering, code review, brainstorming, test generation, and explanations.Last updated -MIT License
- AsecurityAlicenseAqualityA secure Model Context Protocol server that enables Claude Code to connect with OpenAI and Google Gemini models, allowing users to query multiple AI providers through a standardized interface.Last updated -32MIT License
- -securityFlicense-qualityAn enhanced Model Context Protocol server that enables Claude to seamlessly collaborate with multiple AI models (Gemini, OpenAI, local models) for code analysis and development tasks, maintaining context across conversations.Last updated -4533