llm_route
Routes AI prompts to external LLMs based on complexity analysis. Automatically assigns simple queries to budget models, moderate tasks to balanced options, and complex reasoning to premium tiers for cost optimization.
Instructions
Smart router — classifies task complexity, then routes to the optimal external LLM.
Uses a cheap classifier to assess complexity, then picks the right model tier:
- simple → budget models (Gemini Flash, GPT-4o-mini)
- moderate → balanced models (GPT-4o, Sonnet, Gemini Pro)
- complex → premium models (o3, Opus)
For routing to Claude Code's own models (haiku/sonnet) without API keys,
use llm_classify instead and follow its recommendation.
Args:
prompt: The task or question to route.
task_type: Optional hint — "query", "research", "generate", "analyze", "code". Auto-detected if omitted.
complexity_override: Skip classification — force "simple", "moderate", or "complex".
system_prompt: Optional system instructions.
temperature: Sampling temperature (0.0-2.0).
max_tokens: Maximum output tokens.
context: Optional conversation context to help the model understand the broader task.Input Schema
| Name | Required | Description | Default |
|---|---|---|---|
| prompt | Yes | ||
| task_type | No | ||
| complexity_override | No | ||
| system_prompt | No | ||
| temperature | No | ||
| max_tokens | No | ||
| context | No |
Output Schema
| Name | Required | Description | Default |
|---|---|---|---|
| result | Yes |