Integrates with GitHub Copilot CLI to provide access to local LLM capabilities through MCP tools for task delegation and reasoning.
Supports Google Gemini as an optional cloud fallback backend for task delegation and reasoning when local models are unavailable.
Provides integration with Ollama for running local LLMs, including automatic model selection, context-aware routing, and support for multiple model tiers (quick, coder, MoE, thinking).
Delia
A Model Context Protocol (MCP) server that cultivates your local LLM garden. Plant a seed, let Delia pick the right vine, and harvest a fresh melon.
Delia - from Greek Δηλία, "from Delos" (the sacred island). Also, she grows watermelons.
Features
Smart Vine Selection: Routes seeds to the right vine - quick (7B), coder (14B+), moe (30B+), or thinking
Multi-Garden Support: Ollama, llama.cpp, and Gemini gardens with automatic failover
Context-Aware Routing: Handles large seeds with appropriate context windows
Circuit Breaker: Drought protection with graceful recovery
Parallel Processing: Tends multiple seeds simultaneously
Authentication: Optional greenhouse access control
Usage Tracking: Per-gardener quotas and harvest monitoring
Dashboard: Real-time garden status with watermelon-themed activity feed
Requirements
Hardware
Component | Minimum | Recommended | Large Models |
GPU | 4GB VRAM | 12GB VRAM | 24GB+ VRAM |
RAM | 8GB | 16GB | 32GB+ |
Storage | 10GB | 30GB | 50GB+ |
Software
Python 3.11+
uv package manager
One or more backends:
Quick Start
Integration
Delia works with AI coding assistants via MCP. Choose your tool:
VS Code / GitHub Copilot
Add to ~/.config/Code/User/mcp.json:
Reload VS Code to activate.
Claude Code
Create ~/.claude/mcp.json:
Then run claude and use @delia to delegate tasks.
Gemini CLI
Option 1: HTTP Mode (Recommended)
Add to ~/.gemini/settings.json:
Option 2: STDIO Mode
Add to ~/.gemini/settings.json:
GitHub Copilot CLI
Create ~/.copilot-cli/mcp.json:
Configuration
Backend Configuration
Edit settings.json in the project root:
Gemini Cloud Backend (Optional)
Add Gemini as a cloud fallback:
Add to settings.json:
Authentication (Optional)
For HTTP mode with multiple users:
Supports username/password and Microsoft 365 OAuth.
Transport Modes
Tools
Delia provides these MCP tools:
Tool | Description |
| Execute tasks with automatic model selection |
| Extended reasoning for complex problems |
| Process multiple tasks in parallel |
| Check backend status and statistics |
| List available models and tiers |
| Switch between backends at runtime |
| Change model for a tier |
| Get model specifications |
Vine Selection
Delia picks the right vine for every seed:
Vine | Size | Best For |
Quick | 7B-14B | Summaries, simple questions |
Coder | 14B-30B | Generation, review, debugging |
MoE | 30B+ | Architecture, critique, analysis |
Thinking | Specialized | Extended reasoning, research |
Override with hints in your prompt: "use the large model" or "quick answer".
Troubleshooting
Server won't start
MCP not connecting
Verify path in config points to correct directory
Reload VS Code / restart Claude Code
Check logs:
~/.cache/delia/live_logs.json
"Unknown" responses
Backend not running or unreachable
Check
settings.jsonconfigurationRun
curl http://localhost:11434/health
Slow responses
Try smaller models
Check system resources (
nvidia-smi,htop)Reduce context size in
settings.json
Performance
Typical harvest times (modern hardware):
Quick vine: 2-5 seconds
Coder vine: 5-15 seconds
MoE/Thinking vines: 30-60 seconds
License
BSD 3-Clause
Acknowledgments
Ollama - Local LLM runtime
MCP Python SDK - Protocol implementation
Qwen - Base models