Skip to main content
Glama

Server Configuration

Describes the environment variables required to run the server.

NameRequiredDescriptionDefault

No arguments

Capabilities

Server capabilities have not been inspected yet.

Tools

Functions exposed to the LLM to take actions

NameDescription
understand_question

Produce a protocol shell to decompose a user question.

Args: question: The raw user ask to unpack. context: Optional background knowledge or situational frame. constraints: Explicit limits or success criteria. Returns: A structured prompt guiding the model to restate intent, surface constraints, and prepare clarifying questions before acting.
verify_logic

Generate a verification protocol for a reasoning trace.

Args: claim: The headline answer or assertion to validate. reasoning_trace: The supporting chain-of-thought or proof steps. constraints: Optional guardrails (requirements, risk limits). Returns: Structured prompt that audits assumptions, inference steps, and evidence, then proposes patches for any defects.
backtracking

Produce a recursive backtracking scaffold for error correction.

Args: objective: Overall goal to satisfy. failed_step: The step or subgoal that failed. trace: Optional reasoning trace leading to the failure. constraints: Guardrails or requirements to respect. Returns: Structured prompt that rewinds to last stable state, explores alternatives, and proposes a patched plan.
symbolic_abstract

Convert a concrete expression into abstract variables for reasoning.

Args: expression: The raw text or equation to abstract. mapping_hint: Optional guidance for token-to-symbol mapping. goal: Optional downstream task (e.g., simplify, prove, generalize). Returns: Structured prompt that maps tokens to symbols, restates the problem abstractly, and provides a reversible mapping table.
design_context_architecture
Architects a custom context system based on a high-level goal (The Architect). Returns a blueprint of Sutra components (Molecules, Cells, Organs, Thinking Models). Use this when the user wants to build a persistent agent or complex workflow rather than solving a single immediate task. Args: goal: The user's objective (e.g., "Build a writing assistant that learns my style"). constraints: Optional limits (e.g., "Must be lightweight").
get_technique_guide
Returns a guide to available Context Engineering techniques (The Librarian). Use this to discover the best tool for a given task. Args: category: Filter by 'reasoning', 'workflow', 'code', 'project', or 'all'.
analyze_task_complexity
Analyzes a task to recommend the most efficient tool (The Router). Args: task_description: The user's prompt or task.
get_protocol_shell
Returns a Protocol Shell. Can return a specific pre-defined template or a blank shell. Args: name: The name of the protocol (e.g., 'reasoning.systematic') OR a custom name. intent: (Optional) The intent if creating a custom shell.
get_molecular_template

Returns the Python function for creating molecular contexts (Module 02). Use this to programmatically construct few-shot prompts.

get_prompt_program
Returns a functional pseudo-code prompt template (Module 07). Args: program_type: The type of program ('math', 'debate').
get_cell_protocol
Returns a cell protocol template describing memory behaviors. Args: name: Identifier of the cell protocol (key_value, windowed, episodic).
get_organ
Returns an organ template for multi-agent orchestration (Layer 4). Organs combine programs and cells into cohesive workflows for complex tasks requiring multi-perspective analysis or collaborative reasoning. Args: name: Identifier of the organ ('debate_council' for multi-perspective debate).

Prompts

Interactive templates invoked by user choice

NameDescription

No prompts

Resources

Contextual data attached and managed by the client

NameDescription
get_cot_molecules Returns Chain-of-Thought templates (Module 02).
get_reference_layers Returns the Context Engineering Layer definitions.
get_neural_fields Returns Neural Field primitives (Module 08-10).

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/4rgon4ut/sutra'

If you have feedback or need assistance with the MCP directory API, please join our Discord server