Integrates Google's Gemini models (gemini-2.0-flash-exp, gemini-1.5-pro) for AI generation with automatic model selection based on task complexity
Supports local AI models through Ollama integration with custom API endpoints for self-hosted AI generation
Provides access to OpenAI models including GPT-4o, GPT-4o-mini, O1, and O3-mini with intelligent model selection and cost control features
Uses Redis for conversation memory management and persistent storage of conversation context across multi-turn interactions
🧙 SAGE-MCP: Simple AI Guidance Engine for Claude
Universal AI assistant MCP server with intelligent mode selection, conversation continuity, and smart file handling
SAGE-MCP transforms Claude into a multi-talented development assistant that adapts to your needs. Whether you're debugging code, planning architecture, writing tests, or having a technical discussion, SAGE automatically selects the right approach and model for optimal results.
✨ Key Features
🎯 Intelligent Mode System
- chat - Natural conversations with context awareness
- analyze - Deep code analysis and pattern recognition
- review - Comprehensive code reviews with actionable feedback
- debug - Systematic debugging and root cause analysis
- plan - Strategic project planning and architecture design
- test - Test generation with coverage analysis
- refactor - Code improvement and modernization
- think - Deep reasoning with adjustable thinking depth
🔄 Conversation Continuity
- Seamless multi-turn conversations across different modes
- Automatic context preservation between tool calls
- Smart file deduplication - never re-read the same files
- Thread-based memory system for long-running tasks
🤖 Smart Model Selection
- Auto mode - Intelligent model selection based on task complexity
- Support for multiple providers: OpenAI, Anthropic, Google, OpenRouter
- Model restrictions via environment variables for cost control
- Thinking depth control: minimal (0.5%), low (8%), medium (33%), high (67%), max (100%)
📁 Intelligent File Handling
- embedded - Full file content in context (default)
- summary - Token-efficient summaries for large codebases
- reference - File storage with ID references
- Automatic directory expansion and smart deduplication
- Security validation for all file operations
🌐 Web Search Integration
- Real-time documentation lookup
- Best practices and current standards
- Framework and library research
- Error and issue investigation
🎨 Mode Specializations
Mode | Temperature | Description | Best For |
---|---|---|---|
chat | 0.5 | Natural conversations with balanced creativity | Q&A, brainstorming, explanations |
analyze | 0.2 | Focused precision for code analysis | Architecture review, pattern detection |
review | 0.3 | Systematic evaluation with consistent standards | Security audits, best practices |
debug | 0.1 | Deterministic analysis for troubleshooting | Error investigation, root cause analysis |
plan | 0.4 | Strategic thinking for project planning | Architecture design, task breakdown |
test | 0.2 | Accurate test generation with edge cases | Unit tests, integration tests |
refactor | 0.3 | Careful improvements preserving functionality | Code modernization, optimization |
think | 0.7 | Creative problem solving with deep reasoning | Complex algorithms, system design |
🚀 Quick Start
Installation
Claude Desktop Configuration
Add to your Claude Desktop MCP settings:
📖 Usage Examples
Basic Chat
Code Analysis with Files
Multi-turn Conversations
Deep Thinking Mode
Smart File Handling
⚙️ Configuration
Environment Variables
Mode-Specific Temperatures
Default temperatures optimized for each mode:
- chat: 0.5 - Balanced creativity
- analyze: 0.2 - Focused precision
- review: 0.3 - Systematic evaluation
- debug: 0.1 - Deterministic analysis
- plan: 0.4 - Strategic thinking
- test: 0.2 - Accurate test generation
- refactor: 0.3 - Careful improvements
- think: 0.7 - Creative problem solving
🔧 Advanced Features
Conversation Continuation
Smart File Handling
Model Restrictions
Supported Models
Provider | Models | Configuration |
---|---|---|
OpenAI | gpt-4o, gpt-4o-mini, o1, o3-mini | OPENAI_API_KEY |
Anthropic | claude-3-5-sonnet, claude-3-5-haiku | ANTHROPIC_API_KEY |
gemini-2.0-flash-exp, gemini-1.5-pro | GOOGLE_API_KEY | |
OpenRouter | 100+ models from all providers | OPENROUTER_API_KEY |
Custom/Ollama | llama3.2, mistral, codestral | CUSTOM_API_URL |
Complete Configuration Reference
Variable | Description | Example |
---|---|---|
API Keys | ||
OPENAI_API_KEY | OpenAI API key | sk-... |
ANTHROPIC_API_KEY | Anthropic Claude API key | sk-ant-... |
GEMINI_API_KEY / GOOGLE_API_KEY | Google Gemini API key | AIzaSy... |
OPENROUTER_API_KEY | OpenRouter API key | sk-or-... |
XAI_API_KEY | xAI (Grok) API key | xai-... |
CUSTOM_API_URL | Custom/Ollama API endpoint | http://localhost:11434 |
CUSTOM_API_KEY | Custom API key (if required) | custom-key |
Model Selection | ||
DEFAULT_MODEL | Default model (auto for selection) | o3 , gpt-5 , auto |
Model Restrictions | ||
OPENAI_ALLOWED_MODELS | Allowed OpenAI models | o3,gpt-5 |
GOOGLE_ALLOWED_MODELS | Allowed Google models | gemini-2.5-pro,gemini-2.5-flash |
ANTHROPIC_ALLOWED_MODELS | Allowed Anthropic models | claude-3-5-sonnet |
BLOCKED_MODELS | Blocked models (any provider) | gpt-4,o3-mini |
DISABLED_MODEL_PATTERNS | Disable by pattern | anthropic,claude,mini |
Limits & Performance | ||
MAX_FILE_SIZE | Maximum file size in bytes | 5242880 (5MB) |
MCP_PROMPT_SIZE_LIMIT | MCP transport limit | 50000 |
MAX_CONVERSATION_TURNS | Max turns per conversation | 20 |
CONVERSATION_TIMEOUT_HOURS | Conversation timeout | 3 |
Memory & Storage | ||
REDIS_URL | Redis connection for memory | redis://localhost:6379/0 |
REDIS_DB | Redis database number | 0 |
Temperature Overrides | ||
TEMPERATURE_CHAT | Chat mode temperature | 0.7 |
TEMPERATURE_ANALYZE | Analyze mode temperature | 0.3 |
TEMPERATURE_DEBUG | Debug mode temperature | 0.2 |
TEMPERATURE_PLAN | Plan mode temperature | 0.4 |
TEMPERATURE_TEST | Test mode temperature | 0.3 |
TEMPERATURE_REFACTOR | Refactor mode temperature | 0.4 |
TEMPERATURE_REVIEW | Review mode temperature | 0.5 |
TEMPERATURE_THINK | Think mode temperature | 0.8 |
🏗️ Architecture
🧪 Advanced Features
Model Restrictions
Control which models can be used to manage costs:
Conversation Memory
SAGE maintains conversation context across tool calls:
Custom Providers
Add custom AI providers by implementing the base provider interface:
🤝 Contributing
We welcome contributions! Please see CONTRIBUTING.md for guidelines.
Development Setup
📄 License
MIT License - see LICENSE for details.
🙏 Acknowledgments
- Built on FastMCP framework
- Inspired by zen-mcp-server
- Powered by Claude MCP protocol
🔗 Links
SAGE-MCP - Your intelligent AI assistant that adapts to how you work 🧙✨
This server cannot be installed
remote-capable server
The server can be hosted and run remotely because it primarily relies on remote services or has no dependency on the local environment.
A universal AI assistant MCP server that transforms Claude into a multi-talented development assistant with intelligent mode selection, conversation continuity, and smart file handling. Automatically adapts to different tasks like debugging, code analysis, planning, and testing while supporting multiple AI providers and maintaining context across conversations.
Related MCP Servers
- AsecurityAlicenseAqualityAn MCP server that implements Claude Code-like functionality, allowing the AI to analyze codebases, modify files, execute commands, and manage projects through direct file system interactions.Last updated -15248MIT License
- -securityFlicense-qualityAn MCP server that allows AI assistants like Claude to execute terminal commands on the user's computer and return the output, functioning like a terminal through AI.Last updated -52
- AsecurityAlicenseAqualityMCP server that provides Claude AI assistants with the ability to search the web, get news, and perform research using the You.com API.Last updated -4MIT License
- -securityFlicense-qualityA production-ready MCP server ecosystem providing Claude AI with 150+ specialized tools across enhanced memory, data analytics, security, design, and infrastructure domains with PostgreSQL, Redis, Qdrant, and Docker orchestration.Last updated -