Skip to main content
Glama

VARRD — AI Trading Research & Backtesting

Server Details

AI trading research: event studies, backtesting, statistical validation on stocks, futures, crypto.

Status
Healthy
Last Tested
Transport
Streamable HTTP
URL
Repository
augiemazza/varrd
GitHub Stars
13
Server Listing
Varrd

Glama MCP Gateway

Connect through Glama MCP Gateway for full control over tool access and complete visibility into every call.

MCP client
Glama
MCP server

Full call logging

Every tool call is logged with complete inputs and outputs, so you can debug issues and audit what your agents are doing.

Tool access control

Enable or disable individual tools per connector, so you decide what your agents can and cannot do.

Managed credentials

Glama handles OAuth flows, token storage, and automatic rotation, so credentials never expire on your clients.

Usage analytics

See which tools your agents call, how often, and when, so you can understand usage patterns and catch anomalies.

100% free. Your data is private.

Tool Definition Quality

Score is being calculated. Check back soon.

Available Tools

9 tools
autonomous_researchAInspect

Launch VARRD's autonomous research engine to discover and test a trading edge. Give it a topic and it handles everything: generates a creative hypothesis using its concept knowledge base, loads data, charts the pattern, runs the statistical test, and gets the trade setup if an edge is found.

BEST FOR: Exploring a space broadly. The autonomous engine excels at tangential idea generation — give it 'momentum on grains' and it might test wheat seasonal patterns, corn spread reversals, or soybean crush ratio momentum. It propagates from your seed idea into related concepts you might not think of. Great for running many hypotheses at scale.

Returns a complete result — edge/no edge, stats, trade setup. Each call tests ONE hypothesis through the full pipeline. Call again for another idea.

Use 'research' instead when YOU have a specific idea to test and want full control over each step.

ParametersJSON Schema
NameRequiredDescriptionDefault
topicYesResearch topic or trading idea (e.g. 'BTC 240min short setups', 'momentum on grains', 'mean reversion after VIX spikes').
contextNoPrior conversation context — recent user queries to use as research inspiration. Optional.
marketsNoFocus on specific markets (e.g. ['ES', 'NQ']). Omit for VARRD to choose.
test_typeNoType of statistical test. Default: event_study.event_study
search_modeNofocused = stay close to topic. explore = creative freedom. Default: focused.focused
asset_classesNoLimit to specific asset classes. Default: all.
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations indicate mutation (readOnlyHint=false) and external interaction (openWorldHint=true). The description adds valuable behavioral context: it explains the full pipeline (hypothesis generation → data loading → statistical testing), clarifies scope ('Each call tests ONE hypothesis'), and documents the return structure ('edge/no edge, stats, trade setup').

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Structured with clear visual sections (BEST FOR, Returns, explicit sibling comparison). Every sentence serves a distinct purpose: functionality explanation, use case guidance, output specification, or alternative routing. Slightly verbose but information-dense with zero redundancy.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a 6-parameter tool with no output schema, the description adequately compensates by detailing the return value structure and explaining domain-specific concepts (VARRD engine, event study vs backtest). Missing only minor details like error handling or credit consumption (relevant given 'buy_credits' sibling exists).

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100%, establishing baseline 3. The description adds semantic value by illustrating parameter interactions with concrete examples ('momentum on grains' for topic, 'focused' vs 'explore' modes explained via 'tangential idea generation' vs 'seed idea'). This contextual mapping exceeds the schema's mechanical descriptions.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description opens with a precise action statement ('Launch VARRD's autonomous research engine to discover and test a trading edge'). It distinguishes clearly from sibling 'research' via the explicit directive 'Use 'research' instead when YOU have a specific idea...', successfully differentiating autonomous exploration from controlled manual testing.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines5/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Contains explicit 'BEST FOR' section stating when to use ('Exploring a space broadly'). Provides concrete usage pattern examples ('give it 'momentum on grains' and it might test wheat...'). Explicitly names sibling alternative and the condition for choosing it ('when YOU have a specific idea to test').

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

buy_creditsAInspect

Buy credits with USDC on Base. Two modes:

  1. Call without payment_intent_id to get a Stripe deposit address.

  2. Send USDC to that address, then call again with the payment_intent_id to confirm and receive credits. Default $5. Free — no credits consumed to call this.

ParametersJSON Schema
NameRequiredDescriptionDefault
amount_centsNoAmount in cents (default 500 = $5.00). Minimum $5.
payment_intent_idNoStripe PaymentIntent ID from a previous buy_credits call. Pass this after sending USDC to confirm payment and receive credits.
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Adds crucial context beyond annotations: identifies Stripe as processor, Base as network, USDC as currency, and notes 'Free — no credits consumed to call this'. Explains the two-phase commit pattern (deposit address generation vs confirmation) which annotations (readOnlyHint=false, openWorldHint=true) only hint at.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Highly efficient structure: one sentence for purpose, numbered list for modes, final sentence for cost structure. Every clause provides actionable information with zero redundancy.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given no output schema, adequately explains return values implicitly ('get a Stripe deposit address', 'receive credits') and documents the external dependency chain (Stripe → Base). Minor gap: could explicitly document error states or timeout behavior for the pending transaction.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

While schema coverage is 100%, the description adds cross-parameter workflow semantics by framing usage as 'Two modes'—explaining the temporal relationship between amount_cents (applies to both) and payment_intent_id (absence=pending, presence=confirmation). This interaction logic isn't captured in isolated schema field descriptions.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Description states specific verb (buy) + resource (credits) + payment method (USDC on Base). Clearly distinguishes from siblings like check_balance or research tools by focusing on payment acquisition.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines5/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Explicitly documents the two-step workflow: 'Call without payment_intent_id' vs 'call again with the payment_intent_id'. Provides clear temporal sequencing for when to use each parameter state.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

check_balanceA
Read-onlyIdempotent
Inspect

Check your credit balance and see available credit packs. Free — no credits consumed. Call this before heavy operations to ensure you have sufficient credits.

ParametersJSON Schema
NameRequiredDescriptionDefault

No parameters

Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations already indicate read-only, non-destructive, idempotent behavior. The description adds valuable domain-specific context that no credits are consumed and describes what data is returned (balance and available packs), enhancing understanding beyond the structured annotations.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Three sentences efficiently structured: purpose declaration, cost disclosure, and usage timing. Every clause earns its place with no redundancy or filler.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a zero-parameter read operation, the description adequately explains what data is returned despite the lack of an output schema. It covers purpose, cost behavior, and usage timing, though it could specify the exact structure of returned balance information.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Tool accepts zero parameters, meeting the baseline expectation for this dimension. No additional parameter documentation is required or present.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Description clearly states the tool checks credit balance and shows available packs using specific verbs. It effectively distinguishes from siblings like buy_credits (checking vs. purchasing) and research/scan (viewing vs. consuming credits).

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Provides clear temporal guidance ('Call this before heavy operations') and cost context ('Free — no credits consumed'). While it implies prerequisites for expensive operations, it does not explicitly name alternatives or negative conditions.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_briefedA
Read-only
Inspect

Get a personalized market news briefing based on your validated edge library. Profiles your strategies, searches today's news for the instruments and setups you actually trade, and writes a concise digest connecting each headline to your specific book.

Each news item includes a ↳ line tying it to your actual positions and edges (e.g. 'your ES momentum setups', 'your GC mean-reversion edge').

Requires at least 5 strong edges in your library. Costs credits.

ParametersJSON Schema
NameRequiredDescriptionDefault

No parameters

Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Beyond the annotations (readOnly, openWorld, non-idempotent), the description adds critical behavioral context: the credit cost, the minimum edge prerequisite, and the specific output format with ↳ lines connecting news to positions. It also clarifies the temporal nature by mentioning 'today's news', aligning with idempotentHint=false.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is efficiently structured in three distinct paragraphs: core functionality, output format example, and constraints. Every sentence provides unique value without redundancy, and the content is front-loaded with the primary action.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Despite the absence of an output schema, the description adequately describes the return value format (concise digest with ↳ lines linking to edges). Combined with the prerequisite and cost disclosures, it provides sufficient context for invocation, though explicit return type structure would improve it further.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The input schema contains zero parameters. According to scoring rules, zero-parameter tools receive a baseline score of 4, as there are no parameters requiring semantic clarification beyond the schema.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description uses specific verbs ('profiles', 'searches', 'writes') and clearly identifies the resource as a 'personalized market news briefing based on your validated edge library'. It distinguishes itself from generic research siblings by emphasizing personalization tied to the user's specific trading positions and edge library.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides clear prerequisites ('Requires at least 5 strong edges in your library') and cost implications ('Costs credits'), which act as usage constraints. However, it does not explicitly name alternative tools (like 'research' or 'search') for when the user lacks sufficient edges.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_hypothesisA
Read-onlyIdempotent
Inspect

Get full detail for a specific hypothesis/strategy. Returns formula, entry/exit rules, direction, performance metrics (win rate, Sharpe, profit factor, max drawdown), version history, and trade levels. Everything an agent needs to understand and act on a strategy.

ParametersJSON Schema
NameRequiredDescriptionDefault
hypothesis_idYesThe hypothesis ID (from search or scan results).
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no output schema provided, the description adds significant value by exhaustively listing return contents (formula, entry/exit rules, performance metrics, version history). This disclosure of data richness helps the agent anticipate the result structure. The annotations already establish safety (readOnly/idempotent), which the description does not contradict.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Three sentences efficiently structured: purpose declaration, detailed return specification, and value proposition. The final sentence ('Everything an agent needs...') is slightly repetitive but justifies the comprehensiveness of the previous sentence. No significant waste.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the absence of an output schema, the description compensates well by enumerating specific return fields (Sharpe, profit factor, drawdown, etc.). With only one simple parameter and clear annotations, the description fulfills its burden, though it could mention error cases (e.g., invalid ID) for completeness.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100% (the hypothesis_id parameter is fully documented in the schema). The description mentions 'specific hypothesis/strategy' which aligns with the parameter purpose, but does not add syntax, format constraints, or semantic details beyond what the schema already provides. Baseline 3 is appropriate.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the verb (Get) and resource (hypothesis/strategy) and specifies scope as 'full detail' which distinguishes it from presumed summary/list siblings like 'search' or 'get_briefed'. However, it does not explicitly name sibling tools or explicitly contrast with them.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides clear context for when to use the tool: when the agent needs to 'understand and act on a strategy' and requires comprehensive data (formula, metrics, rules). The parameter schema further contextualizes this by noting the ID comes 'from search or scan results', implying a workflow. It lacks explicit 'when not to use' exclusions.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

researchAInspect

Talk to VARRD AI — a quant research system with 15 internal tools. Describe any trading idea in plain language, or ask for specific capabilities like the ELROND expert council, backtesting, or stop-loss optimization.

MULTI-TURN: First call creates a session. Keep calling with the same session_id, following context.next_actions each time.

  1. Your idea -> VARRD charts pattern

  2. 'test it' -> statistical test (event study or backtest)

  3. 'show me the trade setup' -> exact entry/stop/target prices

HYPOTHESIS INTEGRITY (critical): VARRD tests ONE hypothesis at a time — one formula, one setup. Never combine multiple setups into one formula or ask to 'test all' — each idea must be tested as a separate hypothesis for the statistics to be valid. Say 'start a new hypothesis' between ideas to reset cleanly.

  • ALLOWED: Test the SAME setup across multiple markets ('test this on ES, NQ, and CL') — same formula, different data.

  • NOT ALLOWED: Test multiple DIFFERENT formulas/setups at once — each is a separate hypothesis requiring its own chart-test-result cycle. If ELROND council returns 4 setups, test each one separately: chart setup 1 -> test -> results -> 'start new hypothesis' -> chart setup 2 -> etc.

KEY CAPABILITIES you can ask for:

  • 'Use the ELROND council on [market]' -> 8 expert investigators

  • 'Optimize the stop loss and take profit' -> SL/TP grid search

  • 'Test this on ES, NQ, and CL' -> multi-market testing

  • 'Simulate trading this with 1.5 ATR stop' -> backtest with stops

EDGE VERDICTS in context.edge_verdict after testing:

  • STRONG EDGE: Significant vs zero AND vs market baseline

  • MARGINAL: Significant vs zero only (beats nothing, but real signal)

  • PINNED: Significant vs market only (flat returns but different from market)

  • NO EDGE: Neither significant test passed

TERMINAL STATES: Stop when context.has_edge is true (edge found) or false (no edge — valid result). Always read context.next_actions.

ParametersJSON Schema
NameRequiredDescriptionDefault
messageYesYour trading idea, research question, or instruction (e.g. 'test it', 'show trade setup').
session_idNoSession ID from a previous call. Omit to start a new research session.
Behavior5/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Beyond annotations (readOnlyHint:false, idempotentHint:false), the description richly discloses behavioral traits: stateful session management (session_id lifecycle), statistical methodology constraints (one hypothesis at a time), expected context fields (edge_verdict, has_edge, next_actions), and interpretation of edge verdict categories (STRONG EDGE, MARGINAL, PINNED, NO EDGE).

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

While lengthy, the description is well-structured with clear headers (MULTI-TURN, HYPOTHESIS INTEGRITY, KEY CAPABILITIES) and front-loaded purpose. Every section is essential given the complexity of the stateful workflow and lack of output schema. Minor verbosity in the examples section prevents a perfect 5.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness5/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the absence of an output schema, the description comprehensively compensates by explaining exactly what to expect in context variables (edge_verdict types, next_actions, has_edge boolean) and detailing the complete lifecycle from session creation through terminal states.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

With 100% schema coverage, the baseline is met. The description adds valuable semantics beyond the schema: it clarifies the session_id lifecycle (omit on first call, reuse thereafter) and provides concrete message examples ('test it', 'show trade setup', 'start new hypothesis') that illustrate the conversational protocol implied by the multi-turn design.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly defines the tool as an interface to 'VARRD AI — a quant research system' with specific capabilities like ELROND council, backtesting, and pattern charting. It effectively distinguishes from the sibling 'autonomous_research' by emphasizing the interactive, multi-turn conversational nature ('Talk to', 'Your idea ->') and session-based workflow.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines5/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Excellent explicit guidance on when and how to use the tool: it details the multi-turn sequence (1. idea -> 2. test -> 3. trade setup), strictly defines hypothesis integrity constraints (ALLOWED vs NOT ALLOWED testing patterns), instructs to use 'start a new hypothesis' between ideas, and specifies terminal states (stop when context.has_edge is true/false).

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

reset_sessionA
DestructiveIdempotent
Inspect

Kill a broken research session and start fresh. Use this when a session gets stuck, produces errors, or enters a bad state. Free — no credits consumed. After resetting, call research without a session_id to start a new clean session.

ParametersJSON Schema
NameRequiredDescriptionDefault
session_idYesThe session_id to reset.
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations establish the destructive/idempotent profile, while the description adds complementary behavioral context: cost transparency ('Free — no credits consumed'), specificity about error states handled, and integration workflow with sibling tools. Does not contradict annotations.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Four efficient sentences cover purpose, trigger conditions, cost policy, and next-step workflow without redundancy. Information is front-loaded with the destructive action declaration.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness5/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a single-parameter destructive operation with rich annotations, the description is complete. It addresses purpose, prerequisites (broken state), side effects (free of charge), and subsequent workflow integration.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

With 100% schema coverage, the input schema fully documents the session_id parameter. The description validates this by mentioning the parameter indirectly via the post-reset workflow instruction, but does not need to add syntax details beyond the schema.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description opens with the specific action 'Kill a broken research session and start fresh' - a concrete verb-resource combination that clearly distinguishes this cleanup/recovery tool from research-creation siblings like 'research' and 'autonomous_research'.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines5/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Provides explicit trigger conditions ('when a session gets stuck, produces errors, or enters a bad state') and clear post-action workflow ('After resetting, call research without a session_id'), effectively guiding the agent through the full recovery lifecycle.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

scanA
Read-onlyIdempotent
Inspect

Scan your saved strategies against current market data to see what's firing right now. Returns exact dollar entry, stop-loss, and take-profit prices for every active signal. Not a vague directional call — exact trade levels based on the validated statistical model.

ParametersJSON Schema
NameRequiredDescriptionDefault
marketNoFilter by market symbol (e.g. 'ES', 'SPY'). Omit to scan all.
only_firingNoIf true, only return strategies that are firing right now.
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Since annotations declare readOnly/idempotent/safe behavior, the description adds valuable output context: 'exact dollar entry, stop-loss, and take-profit prices' and methodology ('validated statistical model'). This compensates for the missing output schema by clearly stating what data structure users receive.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Three tightly structured sentences cover purpose, output specification, and differentiation. Every sentence earns its place with zero redundancy. Information is front-loaded with the core action in the first sentence.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness5/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given no output schema, the description comprehensively describes returned data (exact dollar levels for entry/stop/profit). Combined with complete parameter documentation and annotations covering side effects, the description provides sufficient context for invocation.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100% with both 'market' and 'only_firing' fully documented. The description does not add parameter syntax or format details beyond the schema, so baseline 3 is appropriate.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Excellent specific verb ('Scan') plus clear resource ('saved strategies') and scope ('firing right now'). Distinguishes from research-focused siblings (research, autonomous_research, get_hypothesis) by emphasizing exact trade execution levels rather than directional analysis or hypotheses.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The phrase 'Not a vague directional call' implicitly guides users to choose this over research tools when exact entry/exit prices are needed. However, it does not explicitly name sibling alternatives or state when to prefer research vs scanning.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

Discussions

No comments yet. Be the first to start the discussion!

Try in Browser

Your Connectors

Sign in to create a connector for this server.