Enables consultation with Google's Gemini models (Gemini 3 Pro, Gemini 2.5 Pro/Flash) via OpenRouter for analyzing large codebases and documents with up to 1M token context windows and advanced reasoning features.
Enables consultation with OpenAI's GPT models (GPT-5.2) via OpenRouter for analyzing extensive file collections with 400k token context windows and reasoning capabilities.
Click on "Install Server".
Wait a few minutes for the server to deploy. Once ready, it will show a "Started" state.
In the chat, type
@followed by the MCP server name and your instructions, e.g., "@Consult7analyze my Python project's src/*.py files for security issues using gemt"
That's it! The server will respond to your query, and you can continue using it as needed.
Here is a step-by-step guide with screenshots.
Consult7 MCP Server
Consult7 is a Model Context Protocol (MCP) server that enables AI agents to consult large context window models via OpenRouter for analyzing extensive file collections - entire codebases, document repositories, or mixed content that exceed the current agent's context limits.
Why Consult7?
Consult7 enables any MCP-compatible agent to offload file analysis to large context models (up to 2M tokens). Useful when:
Agent's current context is full
Task requires specialized model capabilities
Need to analyze large codebases in a single query
Want to compare results from different models
"For Claude Code users, Consult7 is a game changer."
How it works
Consult7 collects files from the specific paths you provide (with optional wildcards in filenames), assembles them into a single context, and sends them to a large context window model along with your query. The result is directly fed back to the agent you are working with.
Example Use Cases
Quick codebase summary
Files:
["/Users/john/project/src/*.py", "/Users/john/project/lib/*.py"]Query: "Summarize the architecture and main components of this Python project"
Model:
"google/gemini-3-flash-preview"Mode:
"fast"
Deep analysis with reasoning
Files:
["/Users/john/webapp/src/*.py", "/Users/john/webapp/auth/*.py", "/Users/john/webapp/api/*.js"]Query: "Analyze the authentication flow across this codebase. Think step by step about security vulnerabilities and suggest improvements"
Model:
"anthropic/claude-sonnet-4.5"Mode:
"think"
Generate a report saved to file
Files:
["/Users/john/project/src/*.py", "/Users/john/project/tests/*.py"]Query: "Generate a comprehensive code review report with architecture analysis, code quality assessment, and improvement recommendations"
Model:
"google/gemini-2.5-pro"Mode:
"think"Output File:
"/Users/john/reports/code_review.md"Result: Returns
"Result has been saved to /Users/john/reports/code_review.md"instead of flooding the agent's context
Featured: Gemini 3 Models
Consult7 supports Google's Gemini 3 family:
Gemini 3 Pro (
google/gemini-3-pro-preview) - Flagship reasoning model, 1M contextGemini 3 Flash (
google/gemini-3-flash-preview) - Ultra-fast model, 1M context
Quick mnemonics for power users:
gemt= Gemini 3 Pro + think (flagship reasoning)gemf= Gemini 3 Flash + fast (ultra fast)gptt= GPT-5.2 + think (latest GPT)grot= Grok 4 + think (alternative reasoning)ULTRA= Run GEMT, GPTT, GROT, and OPUT in parallel (4 frontier models)
These mnemonics make it easy to reference model+mode combinations in your queries.
Installation
Claude Code
Simply run:
Claude Desktop
Add to your Claude Desktop configuration file:
Replace your-openrouter-api-key with your actual OpenRouter API key.
No installation required - uvx automatically downloads and runs consult7 in an isolated environment.
Command Line Options
<api-key>: Required. Your OpenRouter API key--test: Optional. Test the API connection
The model and mode are specified when calling the tool, not at startup.
Supported Models
Consult7 supports all 500+ models available on OpenRouter. Below are the flagship models with optimized dynamic file size limits:
Model | Context | Use Case |
| 400k | Latest GPT, balanced performance |
| 1M | Flagship reasoning model |
| 1M | Best for complex analysis |
| 1M | Gemini 3 Flash, ultra fast |
| 1M | Fast, good for most tasks |
| 1M | Excellent reasoning |
| 200k | Best quality, slower |
| 256k | Alternative reasoning model |
| 2M | Largest context window |
Quick mnemonics:
gptt=openai/gpt-5.2+think(latest GPT, deep reasoning)gemt=google/gemini-3-pro-preview+think(Gemini 3 Pro, flagship reasoning)grot=x-ai/grok-4+think(Grok 4, deep reasoning)oput=anthropic/claude-opus-4.5+think(Claude Opus, deep reasoning)opuf=anthropic/claude-opus-4.5+fast(Claude Opus, no reasoning)gemf=google/gemini-3-flash-preview+fast(Gemini 3 Flash, ultra fast)ULTRA= call GEMT, GPTT, GROT, and OPUT IN PARALLEL (4 frontier models for maximum insight)
You can use any OpenRouter model ID (e.g., deepseek/deepseek-r1-0528). See the full model list. File size limits are automatically calculated based on each model's context window.
Performance Modes
fast: No reasoning - quick answers, simple tasksmid: Moderate reasoning - code reviews, bug analysisthink: Maximum reasoning - security audits, complex refactoring
File Specification Rules
Absolute paths only:
/Users/john/project/src/*.pyWildcards in filenames only:
/Users/john/project/*.py(not in directory paths)Extension required with wildcards:
*.pynot*Mix files and patterns:
["/path/src/*.py", "/path/README.md", "/path/tests/*_test.py"]
Common patterns:
All Python files:
/path/to/dir/*.pyTest files:
/path/to/tests/*_test.pyor/path/to/tests/test_*.pyMultiple extensions:
["/path/*.js", "/path/*.ts"]
Automatically ignored: __pycache__, .env, secrets.py, .DS_Store, .git, node_modules
Size limits: Dynamic based on model context window (e.g., Grok 4 Fast: ~8MB, GPT-5.2: ~1.5MB)
Tool Parameters
The consultation tool accepts the following parameters:
files (required): List of absolute file paths or patterns with wildcards in filenames only
query (required): Your question or instruction for the LLM to process the files
model (required): The LLM model to use (see Supported Models above)
mode (required): Performance mode -
fast,mid, orthinkoutput_file (optional): Absolute path to save the response to a file instead of returning it
If the file exists, it will be saved with
_updatedsuffix (e.g.,report.md→report_updated.md)When specified, returns only:
"Result has been saved to /path/to/file"Useful for generating reports, documentation, or analyses without flooding the agent's context
zdr (optional): Enable Zero Data Retention routing (default:
false)When
true, routes only to endpoints with ZDR policy (prompts not retained by provider)ZDR available: Gemini 3 Pro/Flash, Claude Opus 4.5, GPT-5
Not available: GPT-5.2, Grok 4 (returns error)
Usage Examples
Via MCP in Claude Code
Claude Code will automatically use the tool with proper parameters:
Via Python API
Testing
Uninstalling
To remove consult7 from Claude Code:
Version History
v3.3.0
Fixed GPT-5.2 thinking mode truncation issue (switched to streaming)
Added
google/gemini-3-flash-preview(Gemini 3 Flash, ultra fast)Updated
gemfmnemonic to use Gemini 3 FlashAdded
zdrparameter for Zero Data Retention routing
v3.2.0
Updated to GPT-5.2 with effort-based reasoning
v3.1.0
Added
google/gemini-3-pro-preview(1M context, flagship reasoning model)New mnemonics:
gemt(Gemini 3 Pro),grot(Grok 4),ULTRA(parallel execution)
v3.0.0
Removed Google and OpenAI direct providers - now OpenRouter only
Removed
|thinkingsuffix - usemodeparameter instead (now required)Clean
modeparameter API:fast,mid,thinkSimplified CLI from
consult7 <provider> <key>toconsult7 <key>Better MCP integration with enum validation for modes
Dynamic file size limits based on model context window
v2.1.0
Added
output_fileparameter to save responses to files
v2.0.0
New file list interface with simplified validation
Reduced file size limits to realistic values
License
MIT