Skip to main content
Glama

Server Details

Real-time options analytics: GEX, exposure, greeks, volatility, VRP for US equities

Status
Healthy
Last Tested
Transport
Streamable HTTP
URL
Repository
FlashAlpha-lab/flashalpha-mcp
GitHub Stars
0

Glama MCP Gateway

Connect through Glama MCP Gateway for full control over tool access and complete visibility into every call.

MCP client
Glama
MCP server

Full call logging

Every tool call is logged with complete inputs and outputs, so you can debug issues and audit what your agents are doing.

Tool access control

Enable or disable individual tools per connector, so you decide what your agents can and cannot do.

Managed credentials

Glama handles OAuth flows, token storage, and automatic rotation, so credentials never expire on your clients.

Usage analytics

See which tools your agents call, how often, and when, so you can understand usage patterns and catch anomalies.

100% free. Your data is private.
Tool DescriptionsA

Average 3.8/5 across 37 of 37 tools scored. Lowest: 3.2/5.

Server CoherenceA
Disambiguation4/5

Most tools are clearly distinct, targeting specific aspects of options analytics (e.g., greeks, exposure metrics, volatility, historical replays). However, some overlap exists between 'get_historical_exposure_summary' and individual historical exposure tools (e.g., 'get_historical_gex'), which could cause confusion about which to use for detailed vs. summary data. Descriptions help clarify, but the sheer number of historical tools creates some redundancy.

Naming Consistency5/5

Tool names follow a highly consistent verb_noun pattern throughout, with 'get_' for retrieval, 'calculate_' for computations, and 'solve_' for solving. All use snake_case uniformly, and naming conventions are predictable (e.g., 'get_historical_' prefix for replay tools). This consistency makes the tool set easy to navigate and understand.

Tool Count2/5

With 37 tools, the count is excessive for a single server, leading to potential overwhelm and decision fatigue for agents. While the domain (options analytics) is complex, many tools are highly specialized or redundant (e.g., multiple historical exposure tools), suggesting the surface could be consolidated without losing functionality. This heavy tool count feels bloated and inefficient.

Completeness5/5

The tool set provides comprehensive coverage for options trading analytics, including live and historical data, exposure metrics (GEX, DEX, VEX, CHEX), volatility analysis, greeks calculations, and account management. It supports full CRUD-like workflows for data retrieval and analysis, with no obvious gaps—agents can perform detailed backtesting, real-time monitoring, and strategic calculations seamlessly.

Available Tools

38 tools
calculate_greeksCalculate Option GreeksA
Read-only
Inspect

Calculate Black-Scholes option greeks (delta, gamma, theta, vega, rho, vanna, charm, speed, zomma, color). Pure math — no market data needed.

ParametersJSON Schema
NameRequiredDescriptionDefault
dteYesDays to expiration
spotYesCurrent stock price
typeYes'call' or 'put'
sigmaYesImplied volatility as decimal (0.20 = 20%)
apiKeyYesYour FlashAlpha API key
strikeYesStrike price
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations provide readOnlyHint=true, and the description adds valuable context: it's a 'pure math' calculation that doesn't require external market data, which clarifies the tool's self-contained nature. However, it doesn't mention computational limits, error handling, or output format details.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two sentences with zero waste: the first states the purpose and scope, the second adds critical behavioral context. Every word earns its place, and the information is front-loaded with no unnecessary elaboration.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a calculation tool with read-only annotations and full schema coverage, the description is mostly complete. It clearly explains the tool's mathematical nature and data independence. However, without an output schema, it doesn't describe the return format or structure of the greeks, leaving a minor gap.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, with all 6 parameters well-documented in the schema. The description adds no additional parameter semantics beyond what's already in the schema, so it meets the baseline of 3 for high schema coverage without compensating value.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the specific action ('calculate') and resource ('Black-Scholes option greeks'), listing all 10 greeks by name. It distinguishes itself from siblings by emphasizing 'pure math — no market data needed,' contrasting with data-fetching tools like get_option_quote or get_stock_quote.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines5/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Explicitly states when to use this tool ('no market data needed') and implicitly when not to use it (when live market data is required). The context distinguishes it from siblings that fetch real-time or historical data, providing clear alternatives for different use cases.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

calculate_kellyCalculate Kelly SizingA
Read-only
Inspect

Compute Kelly criterion optimal position sizing for an option trade. Uses BSM expected value vs premium to find edge-maximizing bet size.

ParametersJSON Schema
NameRequiredDescriptionDefault
muYesExpected annual return of underlying as decimal (0.10 = 10%)
dteYesDays to expiration
spotYesCurrent stock price
typeYes'call' or 'put'
sigmaYesImplied volatility as decimal (0.20 = 20%)
apiKeyYesYour FlashAlpha API key
strikeYesStrike price
premiumYesOption premium paid
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations already declare readOnlyHint=true, so the agent knows this is a safe calculation tool. The description adds useful context about the mathematical approach ('Uses BSM expected value vs premium') and objective ('edge-maximizing bet size'), but doesn't disclose computational limits, accuracy constraints, or error handling behavior.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two tightly focused sentences with zero waste. The first sentence states the purpose, the second explains the methodology and objective. Every word earns its place in this efficient description.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a calculation tool with readOnlyHint annotation and comprehensive parameter documentation, the description provides adequate context about what it does and how it works. The main gap is the lack of output schema, but the description compensates somewhat by indicating it returns 'optimal position sizing' and 'edge-maximizing bet size' results.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

With 100% schema description coverage, all 8 parameters are well-documented in the schema. The description adds no additional parameter information beyond what's already in the schema, so it meets the baseline of 3 where the schema does the heavy lifting.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the specific action ('Compute Kelly criterion optimal position sizing') and resource ('for an option trade'), distinguishing it from siblings like 'calculate_greeks' or 'solve_iv' by focusing on position sizing rather than Greeks or implied volatility calculations.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage context ('for an option trade') and mentions the method ('Uses BSM expected value vs premium'), but provides no explicit guidance on when to use this tool versus alternatives like 'calculate_greeks' or other financial calculation tools in the sibling list.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_accountGet Account InfoA
Read-only
Inspect

Get your account info: plan, daily quota limit, usage today, remaining calls.

ParametersJSON Schema
NameRequiredDescriptionDefault
apiKeyYesYour FlashAlpha API key
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations already declare readOnlyHint=true, so the agent knows this is a safe read operation. The description adds useful context about what specific account information is returned (plan, quota, usage), which goes beyond the annotations. However, it doesn't describe rate limits, authentication requirements beyond the API key parameter, or error behavior.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, efficient sentence that front-loads the purpose and lists all returned data points without unnecessary words. Every element serves a clear purpose in informing the user about what the tool provides.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a simple read-only tool with one well-documented parameter and no output schema, the description is reasonably complete. It specifies the exact data returned, which compensates for the lack of output schema. However, it could benefit from mentioning authentication context or typical use cases.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, with the single parameter (apiKey) fully documented in the schema. The description adds no additional parameter information beyond what the schema provides, maintaining the baseline score for high schema coverage.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the specific action ('Get') and resource ('account info'), listing the exact data returned (plan, daily quota limit, usage today, remaining calls). It distinguishes from siblings by focusing on account metadata rather than financial calculations or market data.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides no guidance on when to use this tool versus alternatives. It doesn't mention prerequisites (like needing an API key), nor does it compare with other account-related tools (none exist in siblings, but still lacks context for when this is appropriate).

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_advanced_volatilityGet Advanced VolatilityA
Read-only
Inspect

Get advanced volatility analytics: SVI parameters, forward prices, total variance surface, arbitrage detection, greeks surfaces (vanna, charm, volga, speed), and variance swap fair values. Alpha tier required.

ParametersJSON Schema
NameRequiredDescriptionDefault
apiKeyYesYour FlashAlpha API key
symbolYesStock/ETF ticker
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations already declare readOnlyHint=true, indicating a safe read operation. The description adds value by specifying the 'Alpha tier required' constraint, which is useful context beyond annotations. However, it lacks details on behavioral traits like rate limits, response format, or error handling, leaving gaps in transparency.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is front-loaded with the core purpose in the first clause, followed by a detailed list of outputs and a prerequisite note. It uses two efficient sentences with zero waste, making it appropriately sized and easy to parse.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's complexity (advanced analytics with multiple outputs) and lack of output schema, the description does well by listing specific outputs. However, it could improve by mentioning response structure or limitations, as annotations only cover read-only status, leaving some behavioral aspects undocumented.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, with parameters 'symbol' and 'apiKey' fully documented in the schema. The description does not add any parameter-specific semantics beyond what the schema provides, such as format examples or constraints, so it meets the baseline for high schema coverage.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the specific action ('Get') and resource ('advanced volatility analytics'), listing detailed outputs like SVI parameters, forward prices, total variance surface, arbitrage detection, greeks surfaces, and variance swap fair values. It distinguishes from siblings by specifying 'advanced' analytics, unlike simpler tools like 'get_volatility' or 'get_option_chain'.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description explicitly states 'Alpha tier required', providing clear context on prerequisites. However, it does not specify when to use this tool versus alternatives like 'get_volatility' or 'calculate_greeks', missing explicit guidance on sibling differentiation beyond the 'advanced' qualifier.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_chexGet Charm Exposure (CHEX)B
Read-only
Inspect

Get charm exposure (CHEX) by strike. Shows how dealer delta hedging changes as time passes — reveals time-decay-driven flows.

ParametersJSON Schema
NameRequiredDescriptionDefault
apiKeyYesYour FlashAlpha API key
symbolYesStock/ETF ticker
expirationNoOptional expiration date YYYY-MM-DD
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations provide readOnlyHint=true, which the description doesn't contradict. The description adds valuable behavioral context about what the tool reveals ('how dealer delta hedging changes as time passes — reveals time-decay-driven flows'), which goes beyond the annotations. However, it doesn't mention authentication requirements (apiKey parameter), rate limits, or response format details that would be helpful given no output schema exists.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is extremely concise and front-loaded. The first clause states the core purpose ('Get charm exposure (CHEX) by strike'), and the second clause provides valuable additional context about what the metric reveals. Every word earns its place with zero wasted text or redundancy.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool has good annotations (readOnlyHint) and 100% schema coverage, the description provides adequate purpose explanation. However, with no output schema and multiple similar sibling tools, the description should ideally provide more context about the return format and differentiation from alternatives. The current description is complete enough for basic understanding but leaves gaps in usage guidance.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, so the schema already fully documents all three parameters (symbol, apiKey, expiration). The description doesn't add any parameter-specific information beyond what's in the schema. With complete schema coverage, the baseline score of 3 is appropriate as the description doesn't enhance parameter understanding but doesn't need to compensate for gaps.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool's purpose: 'Get charm exposure (CHEX) by strike' with additional context about what it reveals ('how dealer delta hedging changes as time passes — reveals time-decay-driven flows'). It distinguishes from siblings by focusing on a specific financial metric (CHEX) rather than other calculations like greeks, volatility, or quotes. However, it doesn't explicitly contrast with the most similar sibling 'get_exposure_summary'.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides no guidance on when to use this tool versus alternatives. With multiple sibling tools for financial calculations and data retrieval (e.g., calculate_greeks, get_exposure_summary, get_gex), there's no indication of when CHEX analysis is preferred over other exposure or volatility metrics. The description only explains what the tool does, not when it should be selected.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_dexGet Delta Exposure (DEX)A
Read-only
Inspect

Get delta exposure (DEX) by strike. Shows net dealer delta and directional bias from options hedging.

ParametersJSON Schema
NameRequiredDescriptionDefault
apiKeyYesYour FlashAlpha API key
symbolYesStock/ETF ticker
expirationNoOptional expiration date YYYY-MM-DD
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations already declare readOnlyHint=true, so the agent knows this is a safe read operation. The description adds useful context about what data is returned (net dealer delta and directional bias), but doesn't disclose rate limits, authentication requirements beyond the apiKey parameter, or response format details.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, efficient sentence that front-loads the core purpose ('Get delta exposure (DEX) by strike') followed by clarifying details about what data is shown. Every word earns its place with zero waste.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a read-only tool with good annotations and full parameter documentation, the description provides adequate context about what data is returned. However, without an output schema, it could benefit from more detail about response structure or data format to be fully complete.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

With 100% schema description coverage, the schema already documents all parameters thoroughly. The description doesn't add any parameter-specific information beyond what's in the schema, so it meets the baseline for high schema coverage.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the specific action ('Get delta exposure (DEX) by strike') and the resource (options hedging data), distinguishing it from siblings like 'get_exposure_summary' or 'get_gex' by focusing on net dealer delta and directional bias from options hedging specifically.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage for analyzing options hedging delta exposure, but provides no explicit guidance on when to use this tool versus alternatives like 'get_exposure_summary' or 'get_gex', nor any prerequisites or exclusions.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_exposure_summaryGet Exposure SummaryA
Read-only
Inspect

Get full exposure summary: net GEX/DEX/VEX/CHEX, gamma regime (positive/negative), key levels, hedging estimates, zero-DTE breakdown, top strikes.

ParametersJSON Schema
NameRequiredDescriptionDefault
apiKeyYesYour FlashAlpha API key
symbolYesStock/ETF ticker
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations provide readOnlyHint=true, indicating a safe read operation. The description adds valuable behavioral context beyond annotations by specifying the comprehensive nature of the summary (including gamma regime, key levels, hedging estimates, etc.) and hinting at computational complexity through the detailed output list. It doesn't contradict annotations and provides useful operational insight.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, dense sentence that efficiently lists all key output components without unnecessary words. It's front-loaded with the main purpose ('Get full exposure summary') and follows with specific details, making it highly concise and well-structured for quick comprehension.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a read-only tool with no output schema, the description provides a comprehensive list of return metrics, which compensates well for the missing output schema. It covers the tool's purpose and output scope effectively, though it could benefit from mentioning data sources or update frequency to be fully complete.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, with clear descriptions for both parameters (apiKey and symbol). The description doesn't add any parameter-specific information beyond what's in the schema, such as format examples or constraints. However, given the high schema coverage, a baseline score of 3 is appropriate as the schema adequately documents the parameters.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the specific action ('Get full exposure summary') and lists the exact metrics returned (net GEX/DEX/VEX/CHEX, gamma regime, key levels, hedging estimates, zero-DTE breakdown, top strikes). It distinguishes this tool from siblings like get_gex, get_dex, get_vex, get_chex, and get_zero_dte by indicating it provides a comprehensive summary rather than individual components.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage context by listing the specific exposure metrics, suggesting this tool is for obtaining a consolidated view of market exposure data. However, it doesn't explicitly state when to use this versus alternatives like get_stock_summary or the individual metric tools (e.g., get_gex), nor does it mention prerequisites or exclusions.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_gexGet Gamma Exposure (GEX)A
Read-only
Inspect

Get gamma exposure (GEX) by strike. Shows dealer gamma positioning, gamma flip, call/put walls. Reveals where dealer hedging creates support/resistance.

ParametersJSON Schema
NameRequiredDescriptionDefault
apiKeyYesYour FlashAlpha API key
symbolYesStock/ETF ticker (e.g. SPY, QQQ)
expirationNoOptional expiration date YYYY-MM-DD. Omit for all.
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations provide readOnlyHint=true, indicating a safe read operation. The description adds valuable behavioral context beyond annotations by explaining what the tool reveals ('dealer gamma positioning', 'gamma flip', 'call/put walls', 'where dealer hedging creates support/resistance'), which helps the agent understand the analytical nature of the output. No contradiction with annotations exists.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is perfectly concise with two tightly packed sentences that each earn their place. The first sentence states the core function and key outputs, while the second explains the practical significance. No wasted words, and the most important information is front-loaded.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a read-only tool with full parameter documentation but no output schema, the description provides good contextual completeness by explaining what kind of gamma exposure data is returned and its market significance. However, it doesn't describe the return format or structure, which would be helpful given the absence of an output schema.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

With 100% schema description coverage, the input schema already fully documents all three parameters (symbol, apiKey, expiration). The description doesn't add any parameter-specific semantics beyond what's in the schema, so it meets the baseline expectation without providing extra value.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool's purpose with specific verbs ('Get', 'Shows', 'Reveals') and resources ('gamma exposure (GEX) by strike', 'dealer gamma positioning', 'gamma flip', 'call/put walls', 'dealer hedging support/resistance'). It distinguishes from siblings by focusing specifically on gamma exposure metrics rather than broader calculations like 'calculate_greeks' or data retrieval like 'get_option_chain'.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage for analyzing dealer hedging effects and support/resistance levels, but doesn't explicitly state when to use this tool versus alternatives like 'get_exposure_summary' or 'get_vex'. No explicit when-not-to-use guidance or named alternatives are provided, leaving usage context somewhat implied rather than clearly defined.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_historical_advanced_volatilityGet Historical Advanced VolatilityA
Read-only
Inspect

Replay advanced volatility analytics (SVI parameters, forward prices, total variance surface, arbitrage flags, greek surfaces, variance swap fair values) at any minute since April 2018. EOD-stamped (SVI fits refresh daily). Alpha tier.

ParametersJSON Schema
NameRequiredDescriptionDefault
atYesAs-of timestamp: YYYY-MM-DDTHH:mm:ss (ET) or YYYY-MM-DD
apiKeyYesYour FlashAlpha API key (Alpha tier)
symbolYesStock/ETF ticker
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations declare readOnlyHint=true, so the description adds context: data is EOD-stamped with daily SVI fit refreshes and requires Alpha tier access. No contradictions. Rate limits or other behavioral traits are absent, but for a read-only tool this is acceptable.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two sentences deliver comprehensive scope: first lists analytics and time range, second adds freshness and tier. No redundancy, essential information front-loaded.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the complexity of advanced volatility analytics, the description lists key outputs but lacks explanation of terms (e.g., SVI parameters) and no output schema exists. However, the time range and refresh frequency are explicitly stated, making it fairly complete for an agent with domain knowledge.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100% with all three parameters described (symbol, at, apiKey). The description does not add additional parameter semantics beyond what the schema provides, meeting the baseline of 3.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description specifies the tool replays advanced volatility analytics including specific metrics (SVI parameters, forward prices, total variance surface, arbitrage flags, greek surfaces, variance swap fair values) and the time range (any minute since April 2018). This clearly distinguishes it from simpler historical tools like get_historical_volatility and the current get_advanced_volatility.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies use for historical advanced volatility replay at specific timestamps, but does not explicitly state when not to use or contrast with alternative tools like get_historical_volatility or get_advanced_volatility. However, the sibling tool names provide context, and the 'replay' and 'any minute' phrasing suggests time-specific queries.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_historical_chexGet Historical CHEXB
Read-only
Inspect

Replay charm exposure (CHEX) by strike at any minute since April 2018. Alpha tier.

ParametersJSON Schema
NameRequiredDescriptionDefault
atYesAs-of timestamp: YYYY-MM-DDTHH:mm:ss (ET) or YYYY-MM-DD
apiKeyYesYour FlashAlpha API key (Alpha tier)
symbolYesStock/ETF ticker
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations already declare readOnlyHint=true, indicating a safe read operation. The description adds some behavioral context: it specifies the data's temporal scope ('since April 2018') and granularity ('at any minute'), and notes the 'Alpha tier' requirement. However, it lacks details on rate limits, data format, or potential errors, which would be valuable given no output schema.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is concise and front-loaded with the core purpose in the first sentence. The second sentence ('Alpha tier.') adds necessary context efficiently. There's no wasted text, though it could be slightly more structured (e.g., separating usage notes).

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's complexity (historical financial data retrieval), annotations cover safety (readOnlyHint), and schema covers inputs fully. However, with no output schema, the description doesn't explain return values (e.g., data format, structure), which is a gap. It's adequate but incomplete for guiding an agent on what to expect from the tool.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, so the schema fully documents parameters (symbol, apiKey, at). The description doesn't add any parameter-specific semantics beyond what's in the schema, such as explaining 'symbol' beyond 'Stock/ETF ticker' or clarifying 'at' formats. With high schema coverage, a baseline score of 3 is appropriate.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool's purpose: 'Replay charm exposure (CHEX) by strike at any minute since April 2018.' It specifies the verb ('Replay'), resource ('charm exposure'), and temporal scope. However, it doesn't explicitly differentiate from sibling tools like 'get_chex' (presumably current CHEX) or 'get_historical_exposure_summary', which is needed for a perfect score.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides minimal usage guidance. It mentions 'Alpha tier' and implies historical data retrieval, but offers no explicit when-to-use rules, alternatives (e.g., vs. 'get_chex' for current data), or prerequisites beyond the API key. This leaves the agent with little context for tool selection among siblings.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_historical_coverageGet Historical CoverageA
Read-only
Inspect

List symbols backfilled in the historical archive with coverage windows, day counts, and gaps. Call this first to check whether a symbol + date range is queryable before sending a replay request. Alpha tier.

ParametersJSON Schema
NameRequiredDescriptionDefault
apiKeyYesYour FlashAlpha API key (Alpha tier)
symbolNoOptional symbol filter (e.g. SPY) - omit for all covered symbols
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations provide readOnlyHint=true, but the description adds valuable context: it discloses this is for checking coverage metadata (not fetching data), mentions the 'Alpha tier' limitation (implying potential rate limits or access constraints), and clarifies it's a pre-check tool. No contradiction with annotations exists.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two sentences with zero waste: first defines purpose and output, second provides critical usage guidance and tier info. Front-loaded with core functionality, efficiently structured.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a read-only tool with full schema coverage and no output schema, the description is nearly complete. It covers purpose, usage context, and behavioral constraints. Minor gap: no explicit mention of response format (e.g., list structure), but annotations and context mitigate this.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, so the schema fully documents both parameters. The description adds no additional parameter details beyond implying 'symbol' filtering for coverage checks. Baseline 3 is appropriate as the schema carries the burden.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the verb ('List') and resource ('symbols backfilled in the historical archive') with specific attributes ('coverage windows, day counts, and gaps'). It distinguishes from siblings by focusing on coverage metadata rather than actual historical data retrieval like 'get_historical_stock_quote' or calculations like 'calculate_greeks'.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines5/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Explicit guidance is provided: 'Call this first to check whether a symbol + date range is queryable before sending a replay request.' This tells the agent when to use this tool (pre-check) versus alternatives (actual data retrieval tools), with a clear prerequisite workflow.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_historical_dexGet Historical DEXA
Read-only
Inspect

Replay delta exposure (DEX) by strike at any minute since April 2018. Alpha tier.

ParametersJSON Schema
NameRequiredDescriptionDefault
atYesAs-of timestamp: YYYY-MM-DDTHH:mm:ss (ET) or YYYY-MM-DD
apiKeyYesYour FlashAlpha API key (Alpha tier)
symbolYesStock/ETF ticker
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations already declare readOnlyHint=true, indicating a safe read operation. The description adds context about the data availability ('since April 2018') and access tier ('Alpha tier'), which are useful behavioral traits not covered by annotations. However, it lacks details on rate limits, error handling, or response format, leaving some gaps in behavioral understanding.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is highly concise and front-loaded, using only two sentences that efficiently convey purpose, scope, and access tier without unnecessary details. Every sentence earns its place by providing critical information, making it easy to scan and understand.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's complexity (historical data retrieval with three parameters), rich annotations (readOnlyHint), and high schema coverage, the description is mostly complete. It covers purpose, temporal scope, and access tier. However, without an output schema, it does not explain return values (e.g., data format), which is a minor gap in completeness for this context.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, with all parameters well-documented in the input schema (e.g., 'at' as timestamp, 'apiKey' for Alpha tier, 'symbol' as ticker). The description does not add any meaningful parameter semantics beyond what the schema provides, such as clarifying 'symbol' formats or 'at' precision. Baseline 3 is appropriate given the comprehensive schema.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool's purpose with specific verbs ('Replay delta exposure (DEX) by strike') and resources ('at any minute since April 2018'), distinguishing it from siblings like 'get_dex' (likely current data) and 'get_historical_chex' (different metric). It precisely defines the action and temporal scope.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides clear context for usage ('Alpha tier', 'since April 2018'), implying this is for historical data retrieval within a specific timeframe. However, it does not explicitly state when to use this versus alternatives like 'get_dex' (for current data) or other historical tools, nor does it mention exclusions or prerequisites beyond the Alpha tier.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_historical_exposure_summaryGet Historical Exposure SummaryA
Read-only
Inspect

Replay the full exposure summary (net GEX/DEX/VEX/CHEX, regime, hedging estimates, top strikes) at any minute since April 2018. Alpha tier.

ParametersJSON Schema
NameRequiredDescriptionDefault
atYesAs-of timestamp: YYYY-MM-DDTHH:mm:ss (ET) or YYYY-MM-DD
apiKeyYesYour FlashAlpha API key (Alpha tier)
symbolYesStock/ETF ticker
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations provide readOnlyHint=true, indicating this is a safe read operation. The description adds valuable behavioral context beyond annotations: it specifies the historical data availability ('since April 2018'), mentions the 'Alpha tier' requirement (implying authentication and potentially rate limits or access restrictions), and details the comprehensive nature of the summary (including multiple metrics like regime and hedging estimates). No contradiction with annotations exists.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is extremely concise and front-loaded, consisting of a single sentence that efficiently conveys the tool's purpose, scope, and key details (historical range, data components, tier requirement). Every word earns its place with no wasted information, making it easy for an AI agent to parse quickly.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's complexity (historical data retrieval with multiple metrics) and the absence of an output schema, the description does well by specifying what data is returned (net GEX/DEX/VEX/CHEX, regime, hedging estimates, top strikes). However, it doesn't detail the format or structure of the return values, which could be helpful for an AI agent to interpret results. With annotations covering safety and schema covering inputs, it's mostly complete but has a minor gap in output clarification.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, with clear descriptions for all three parameters (symbol, apiKey, at). The description adds minimal parameter semantics beyond the schema, only implying that 'at' corresponds to historical timestamps and 'Alpha tier' relates to apiKey. Since the schema already documents parameters thoroughly, the baseline score of 3 is appropriate, as the description doesn't significantly enhance parameter understanding.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the specific action ('Replay the full exposure summary') and resource ('at any minute since April 2018'), with explicit details about what data is included (net GEX/DEX/VEX/CHEX, regime, hedging estimates, top strikes). It effectively distinguishes itself from sibling tools like get_exposure_summary (which likely provides current data) and other historical tools by specifying the comprehensive exposure summary scope.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides clear context for when to use this tool: for retrieving historical exposure summaries at specific timestamps since April 2018. It mentions 'Alpha tier' which implies access requirements. However, it doesn't explicitly state when not to use it or name specific alternatives among the many sibling tools, such as get_exposure_summary for current data or other historical tools for partial metrics.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_historical_gexGet Historical GEXA
Read-only
Inspect

Replay gamma exposure (GEX) by strike at any minute since April 2018. Returns same shape as live /v1/exposure/gex. Alpha tier.

ParametersJSON Schema
NameRequiredDescriptionDefault
atYesAs-of timestamp: YYYY-MM-DDTHH:mm:ss (ET) or YYYY-MM-DD (defaults to 16:00 ET)
apiKeyYesYour FlashAlpha API key (Alpha tier)
symbolYesStock/ETF ticker (e.g. SPY)
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations already declare readOnlyHint=true, so the agent knows this is a safe read operation. The description adds useful context about the return format ('same shape as live /v1/exposure/gex') and tier restrictions ('Alpha tier'), but doesn't disclose rate limits, authentication requirements beyond the apiKey parameter, or error behaviors.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is perfectly concise with two sentences that each earn their place: the first defines purpose and scope, the second specifies return format and tier. No wasted words, front-loaded with core functionality.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a read-only tool with full parameter documentation and no output schema, the description provides good context about historical scope, return format, and tier requirements. However, it could better address authentication implications of the 'Alpha tier' note and potential limitations of historical data availability.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

With 100% schema description coverage, the input schema fully documents all three parameters. The description doesn't add any parameter-specific information beyond what's in the schema, so it meets the baseline expectation without providing additional semantic context.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the specific action ('Replay gamma exposure (GEX) by strike'), resource ('at any minute since April 2018'), and distinguishes it from sibling 'get_gex' by specifying historical data retrieval. It precisely defines the tool's function with temporal scope.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides clear context for when to use this tool ('Replay gamma exposure... at any minute since April 2018') and implicitly distinguishes it from live GEX tools. However, it doesn't explicitly mention when NOT to use it or name specific alternatives among the many sibling tools.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_historical_levelsGet Historical Key LevelsA
Read-only
Inspect

Replay key options levels (gamma flip, call/put walls, highest OI strike, 0DTE magnet) at any minute since April 2018. Alpha tier.

ParametersJSON Schema
NameRequiredDescriptionDefault
atYesAs-of timestamp: YYYY-MM-DDTHH:mm:ss (ET) or YYYY-MM-DD
apiKeyYesYour FlashAlpha API key (Alpha tier)
symbolYesStock/ETF ticker
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations provide readOnlyHint=true, indicating a safe read operation. The description adds value by specifying the historical timeframe ('since April 2018'), data granularity ('at any minute'), and tier requirement ('Alpha tier'), which are not covered by annotations. However, it lacks details on rate limits, authentication needs beyond apiKey, or output format, leaving some behavioral gaps.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, efficient sentence that front-loads key information (action, resource, timeframe, tier) with zero wasted words. It appropriately sizes the information to the tool's complexity, making it easy to parse quickly.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's complexity (historical data retrieval with specific metrics), annotations cover safety (readOnlyHint), and schema fully documents parameters. However, without an output schema, the description does not explain return values (e.g., format of replayed levels), which is a minor gap. It compensates by specifying the historical scope and tier, making it mostly complete for agent use.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, with clear parameter descriptions in the schema (e.g., timestamp format, API key tier, ticker). The description does not add any additional parameter semantics beyond what the schema provides, such as explaining the significance of the 'at' parameter for historical replay. Baseline score of 3 is appropriate as the schema carries the full burden.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the specific action ('Replay key options levels') and resource ('at any minute since April 2018'), distinguishing it from siblings like 'get_levels' (likely current levels) and other historical tools by specifying the exact metrics (gamma flip, call/put walls, highest OI strike, 0DTE magnet). It provides a precise verb+resource combination with explicit scope.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage context by specifying 'historical' data 'since April 2018' and 'Alpha tier', suggesting this is for historical analysis with premium access. However, it does not explicitly state when to use this tool versus alternatives like 'get_historical_option_quote' or 'get_levels', nor does it provide exclusions or direct comparisons to sibling tools.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_historical_max_painGet Historical Max PainB
Read-only
Inspect

Replay max pain, pain curve, dealer alignment, and pin probability at any minute since April 2018. Alpha tier.

ParametersJSON Schema
NameRequiredDescriptionDefault
atYesAs-of timestamp: YYYY-MM-DDTHH:mm:ss (ET) or YYYY-MM-DD
apiKeyYesYour FlashAlpha API key (Alpha tier)
symbolYesStock/ETF ticker
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations already declare readOnlyHint=true, indicating a safe read operation. The description adds context about the temporal scope ('since April 2018') and access tier ('Alpha tier'), which are useful behavioral details not covered by annotations. However, it doesn't disclose rate limits, data freshness, or response format.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, efficient sentence that front-loads the core functionality. It could be slightly more structured by separating scope from tier information, but it avoids redundancy and wastes no words.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's complexity (historical financial data retrieval), annotations cover safety (read-only), but no output schema exists. The description provides temporal scope and tier context, but lacks details on response structure, data granularity, or error handling, leaving gaps for an AI agent.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, so the schema fully documents all three parameters. The description doesn't add any parameter-specific semantics beyond what's in the schema (e.g., it doesn't explain 'symbol' beyond ticker or 'at' format details). Baseline 3 is appropriate when schema does the heavy lifting.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool 'replays' multiple financial metrics (max pain, pain curve, dealer alignment, pin probability) with a specific temporal scope ('any minute since April 2018'). It distinguishes from siblings by focusing on historical max pain data rather than calculations, quotes, or other metrics, though it doesn't explicitly name alternatives.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage for retrieving historical max pain data, but provides no explicit guidance on when to use this tool versus alternatives like 'get_historical_volatility' or 'calculate_greeks'. The 'Alpha tier' mention hints at access requirements but doesn't clarify functional alternatives.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_historical_narrativeGet Historical NarrativeA
Read-only
Inspect

Replay the verbal narrative analysis (regime, key-level commentary, prior-day comparison) at any minute since April 2018. Alpha tier.

ParametersJSON Schema
NameRequiredDescriptionDefault
atYesAs-of timestamp: YYYY-MM-DDTHH:mm:ss (ET) or YYYY-MM-DD
apiKeyYesYour FlashAlpha API key (Alpha tier)
symbolYesStock/ETF ticker
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations already declare readOnlyHint=true, indicating a safe read operation. The description adds valuable context beyond annotations: it specifies the content type ('verbal narrative analysis'), historical scope ('since April 2018'), and tier requirement ('Alpha tier'), though it doesn't detail rate limits or response format.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is front-loaded with essential information in a single, efficient sentence. Every phrase ('Replay the verbal narrative analysis', 'at any minute since April 2018', 'Alpha tier') adds value without redundancy, making it appropriately sized.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's complexity (historical data retrieval), annotations cover safety, and schema fully documents inputs, the description provides good context on content and scope. However, without an output schema, it doesn't explain return values, leaving a minor gap for a read operation.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, so parameters are fully documented in the schema. The description adds no additional parameter semantics beyond implying historical retrieval via 'at', aligning with the baseline score when schema handles documentation.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the specific action ('Replay') and resource ('verbal narrative analysis') with detailed scope ('regime, key-level commentary, prior-day comparison') and temporal range ('any minute since April 2018'). It distinguishes from siblings like 'get_narrative' by specifying historical retrieval.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage for historical narrative retrieval but provides no explicit guidance on when to use this versus alternatives like 'get_narrative' (current) or other historical tools. It mentions 'Alpha tier' as a prerequisite but doesn't clarify exclusion scenarios or direct comparisons.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_historical_option_quoteGet Historical Option QuoteA
Read-only
Inspect

Replay the full option chain with BSM greeks, IV, OI at any minute since April 2018. Filter by expiry, strike, and type. Alpha tier.

ParametersJSON Schema
NameRequiredDescriptionDefault
atYesAs-of timestamp: YYYY-MM-DDTHH:mm:ss (ET) or YYYY-MM-DD
typeNoOptional 'C' or 'P' (call or put)
apiKeyYesYour FlashAlpha API key (Alpha tier)
expiryNoOptional expiration date YYYY-MM-DD
strikeNoOptional strike price
symbolYesUnderlying ticker
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations provide readOnlyHint=true, indicating safe read operations. The description adds value by specifying the data source ('ClickHouse tick archive'), which gives context about data freshness and reliability. However, it doesn't disclose behavioral traits like rate limits, authentication requirements beyond the apiKey parameter, or error handling. No contradiction with annotations exists.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, efficient sentence that front-loads the core purpose ('Get historical option quote from a specific date') and includes key details (filtering, data source) without redundancy. Every word earns its place, making it easy to parse quickly.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's moderate complexity (7 parameters, historical data retrieval) and annotations covering read-only safety, the description is adequate but has gaps. It lacks output details (no output schema provided), doesn't explain error cases or data availability, and could better differentiate from siblings. For a read-only tool with good schema coverage, it meets minimum viability but isn't fully comprehensive.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, so all parameters are documented in the schema. The description mentions filtering by 'expiry, strike, and type', which aligns with optional parameters in the schema but doesn't add meaning beyond what's already there (e.g., format details or constraints). With high schema coverage, the baseline score of 3 is appropriate as the description doesn't significantly enhance parameter understanding.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool's purpose: 'Get historical option quote from a specific date' with specific filtering capabilities (expiry, strike, type) and data source (ClickHouse tick archive). It distinguishes from siblings like 'get_option_quote' (likely current quotes) and 'get_historical_stock_quote' (stock vs. option), though not explicitly. The verb+resource combination is specific but could be more explicit about sibling differentiation.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage context by specifying 'historical' and filtering parameters, suggesting it's for retrieving past option data rather than current quotes. However, it doesn't explicitly state when to use this vs. alternatives like 'get_option_quote' or 'get_historical_stock_quote', nor does it mention prerequisites or exclusions. The guidance is present but not comprehensive.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_historical_stock_quoteGet Historical Stock QuoteA
Read-only
Inspect

Replay a stock bid/ask/mid at any minute since April 2018. Alpha tier.

ParametersJSON Schema
NameRequiredDescriptionDefault
atYesAs-of timestamp: YYYY-MM-DDTHH:mm:ss (ET) or YYYY-MM-DD
apiKeyYesYour FlashAlpha API key (Alpha tier)
symbolYesStock ticker
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations already declare readOnlyHint=true, indicating a safe read operation. The description adds useful context about the data source ('ClickHouse tick archive') and clarifies that omitting time returns full-day data, but doesn't disclose rate limits, error conditions, or response format. No contradiction with annotations exists.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is two concise sentences with zero waste. The first sentence states the core purpose, and the second adds essential context about the data source. It's front-loaded and efficiently structured.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's moderate complexity (historical data retrieval), annotations cover safety (readOnlyHint), and schema fully documents parameters, but no output schema exists. The description lacks details on return values (e.g., quote fields), error handling, or API limitations, leaving gaps for an agent.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, so parameters are fully documented in the schema. The description adds minimal value beyond the schema, only implying date/time usage without detailing format or constraints. Baseline 3 is appropriate as the schema carries the burden.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the specific action ('Get historical stock quote') and resource ('from a specific date and time'), distinguishing it from siblings like 'get_stock_quote' (current quote) and 'get_historical_option_quote' (options data). It specifies the data source ('ClickHouse tick archive'), adding precision.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage for historical stock data at a specific date/time, but doesn't explicitly state when to use this vs. alternatives like 'get_stock_quote' for current data or 'get_historical_option_quote' for options. No guidance on prerequisites (e.g., API key setup) or exclusions is provided.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_historical_stock_summaryGet Historical Stock SummaryA
Read-only
Inspect

Replay the comprehensive stock summary (price, IV, VRP, exposure, flow, macro) at any minute since April 2018. Alpha tier.

ParametersJSON Schema
NameRequiredDescriptionDefault
atYesAs-of timestamp: YYYY-MM-DDTHH:mm:ss (ET) or YYYY-MM-DD
apiKeyYesYour FlashAlpha API key (Alpha tier)
symbolYesStock/ETF/index ticker
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations provide readOnlyHint=true, indicating a safe read operation. The description adds value by specifying the historical scope ('since April 2018'), data comprehensiveness, and tier requirement ('Alpha tier'), but does not disclose additional behavioral traits like rate limits, response format, or pagination. No contradiction with annotations.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, efficient sentence that front-loads key information (action, resource, scope, data fields, tier). Every word earns its place with no redundancy or waste, making it highly concise and well-structured.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's complexity (historical data retrieval with multiple data fields), annotations cover safety (read-only), and schema fully documents parameters. However, no output schema exists, and the description does not explain return values or format. It is mostly complete but could benefit from output details.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, with clear parameter descriptions in the input schema. The description does not add meaning beyond the schema, such as explaining parameter interactions or constraints. Baseline score of 3 is appropriate as the schema handles parameter documentation adequately.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the action ('Replay'), resource ('comprehensive stock summary'), and scope ('at any minute since April 2018'), with specific data fields listed (price, IV, VRP, exposure, flow, macro). It distinguishes from siblings like get_stock_summary by specifying historical retrieval and the 'Alpha tier' requirement.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage for historical data retrieval since April 2018 and mentions 'Alpha tier' as a prerequisite, but does not explicitly state when to use this tool versus alternatives like get_historical_stock_quote or get_stock_summary. It provides clear context but lacks explicit exclusions or named alternatives.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_historical_surfaceGet Historical IV SurfaceA
Read-only
Inspect

Replay the implied volatility surface grid at any minute since April 2018. EOD-stamped (SVI parameters refresh daily). Alpha tier.

ParametersJSON Schema
NameRequiredDescriptionDefault
atYesAs-of timestamp: YYYY-MM-DDTHH:mm:ss (ET) or YYYY-MM-DD
apiKeyYesYour FlashAlpha API key (Alpha tier)
symbolYesStock/ETF ticker
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations provide readOnlyHint=true, indicating a safe read operation. The description adds valuable behavioral context: it specifies the data availability ('since April 2018'), refresh frequency ('EOD-stamped... refresh daily'), and access tier ('Alpha tier'), which aren't covered by annotations. It doesn't contradict annotations, as 'replay' aligns with read-only behavior. However, it lacks details on rate limits, error handling, or output format.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is highly concise and well-structured in a single sentence: it front-loads the core purpose ('Replay the implied volatility surface grid'), specifies key constraints ('at any minute since April 2018'), adds important context ('EOD-stamped... refresh daily'), and notes access level ('Alpha tier'). Every part earns its place without redundancy.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's complexity (historical data retrieval with temporal precision), annotations cover safety (readOnlyHint), and schema fully documents parameters. The description adds context on data range, refresh behavior, and tier, which is valuable. However, without an output schema, it doesn't describe the return format (e.g., grid structure, SVI parameters), leaving a gap for agent invocation. It's mostly complete but could benefit from output details.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, with clear parameter descriptions in the schema (e.g., 'As-of timestamp: YYYY-MM-DDTHH:mm:ss (ET) or YYYY-MM-DD'). The description doesn't add semantic details beyond the schema, such as explaining 'symbol' beyond 'Stock/ETF ticker' or 'apiKey' usage. With high schema coverage, the baseline score of 3 is appropriate, as the description doesn't compensate but doesn't need to heavily.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool's purpose: 'Replay the implied volatility surface grid at any minute since April 2018.' It specifies the exact resource (implied volatility surface), action (replay), and temporal scope (since April 2018). The mention of 'EOD-stamped (SVI parameters refresh daily)' further distinguishes it from real-time or continuously updating tools, and 'Alpha tier' clarifies access level.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage context by stating 'at any minute since April 2018' and 'EOD-stamped,' suggesting it's for historical data retrieval rather than real-time. However, it doesn't explicitly guide when to use this tool versus alternatives like 'get_historical_volatility' or 'solve_iv' from the sibling list. No exclusions or prerequisites are mentioned beyond the Alpha tier note.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_historical_vexGet Historical VEXB
Read-only
Inspect

Replay vanna exposure (VEX) by strike at any minute since April 2018. Alpha tier.

ParametersJSON Schema
NameRequiredDescriptionDefault
atYesAs-of timestamp: YYYY-MM-DDTHH:mm:ss (ET) or YYYY-MM-DD
apiKeyYesYour FlashAlpha API key (Alpha tier)
symbolYesStock/ETF ticker
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations already declare readOnlyHint=true, indicating a safe read operation. The description adds some behavioral context: it specifies the data availability ('since April 2018') and granularity ('at any minute'), and notes the 'Alpha tier' requirement. However, it doesn't disclose other traits like rate limits, authentication needs beyond the apiKey parameter, or response format. No contradiction with annotations exists.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is highly concise and front-loaded: a single sentence conveys the core functionality, followed by a brief tier note. Every word earns its place with no redundancy or fluff, making it efficient for an agent to parse.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's moderate complexity (historical data retrieval with three parameters), annotations cover safety (readOnlyHint), and schema coverage is complete, the description is adequate but has gaps. It lacks output details (no output schema provided), doesn't explain data format or limitations, and offers minimal sibling differentiation. It meets minimum viability but could be more comprehensive.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, so the input schema fully documents all three parameters (symbol, apiKey, at). The description adds no additional parameter semantics beyond what's in the schema. It mentions 'Alpha tier' which relates to apiKey but doesn't elaborate. Baseline 3 is appropriate as the schema carries the burden.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool's purpose: 'Replay vanna exposure (VEX) by strike at any minute since April 2018.' It specifies the verb ('replay'), resource ('vanna exposure'), and temporal scope. However, it doesn't explicitly differentiate from sibling tools like 'get_vex' or 'get_historical_exposure_summary', which likely serve related but distinct purposes.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides minimal usage guidance. It mentions 'Alpha tier' and implies historical data retrieval, but offers no explicit advice on when to use this tool versus alternatives (e.g., 'get_vex' for current data or other historical tools). There's no mention of prerequisites, exclusions, or comparative context with siblings.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_historical_volatilityGet Historical VolatilityA
Read-only
Inspect

Replay volatility analytics (ATM IV, realised vol, IV-RV spreads, skew, term structure) at any minute since April 2018. Alpha tier.

ParametersJSON Schema
NameRequiredDescriptionDefault
atYesAs-of timestamp: YYYY-MM-DDTHH:mm:ss (ET) or YYYY-MM-DD
apiKeyYesYour FlashAlpha API key (Alpha tier)
symbolYesStock/ETF ticker
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations provide readOnlyHint=true, and the description adds valuable context: it specifies the data availability timeframe ('since April 2018'), mentions the 'Alpha tier' indicating premium/restricted access, and implies replay functionality ('Replay volatility analytics'). This goes beyond annotations by detailing temporal limits and access requirements without contradictions.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, dense sentence with zero waste: it front-loads the core purpose, specifies analytics types, temporal scope, and tier info efficiently. Every element earns its place without redundancy or fluff.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given no output schema, the description could better explain return values (e.g., format of volatility analytics). However, with rich annotations (readOnlyHint) and high schema coverage, it provides good context on purpose, behavior, and constraints. It's mostly complete but lacks output details.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, with clear parameter descriptions in the schema. The description doesn't add specific parameter semantics beyond implying 'at' is for historical timestamps and 'apiKey' is for Alpha tier access, which aligns with schema info. Baseline 3 is appropriate as the schema handles parameter documentation adequately.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the action ('Replay volatility analytics') and specific resources (ATM IV, realised vol, IV-RV spreads, skew, term structure), with temporal scope ('at any minute since April 2018') and tier qualification ('Alpha tier'). It distinguishes from siblings by focusing on historical volatility analytics rather than current calculations or other data types.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage for historical volatility data retrieval ('at any minute since April 2018') but doesn't explicitly state when to use this versus alternatives like 'get_volatility' (likely current) or 'get_historical_surface' (different analytics). The Alpha tier mention suggests access restrictions but no clear guidance on tool selection.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_historical_vrpGet Historical VRPA
Read-only
Inspect

Replay VRP dashboard (z-score, percentile, regime, strategy scores) at any minute since April 2018. Percentiles and z-scores are leak-free: date-bounded in SQL so the backtest only sees data strictly before the at timestamp. Alpha tier.

ParametersJSON Schema
NameRequiredDescriptionDefault
atYesAs-of timestamp: YYYY-MM-DDTHH:mm:ss (ET) or YYYY-MM-DD
apiKeyYesYour FlashAlpha API key (Alpha tier)
symbolYesStock/ETF ticker
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations provide readOnlyHint=true, and the description adds valuable context beyond this: it specifies the historical scope ('since April 2018'), the leak-free nature of percentiles and z-scores via SQL date-bounding, and the Alpha tier requirement. This enhances transparency without contradicting annotations.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is front-loaded with key information in a single, efficient sentence. Every part earns its place by specifying the tool's function, historical scope, data integrity, and tier, with no wasted words.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's complexity (historical data with leak-free calculations), annotations cover safety (read-only), and schema covers parameters fully. However, without an output schema, the description could better explain return values (e.g., structure of VRP dashboard data). It's mostly complete but has a minor gap in output details.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, so the schema fully documents parameters. The description adds no specific parameter semantics beyond what the schema provides, such as format details for 'at' or usage of 'apiKey'. Baseline 3 is appropriate as the schema handles parameter documentation.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool's purpose with specific verbs ('Replay VRP dashboard') and resources ('z-score, percentile, regime, strategy scores'), and distinguishes it from siblings by specifying historical data 'since April 2018' and 'leak-free' backtesting, unlike other tools like 'get_vrp' or 'get_vrp_history'.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides clear context for usage ('at any minute since April 2018', 'leak-free', 'Alpha tier'), but does not explicitly state when not to use it or name alternatives among siblings. It implies usage for historical VRP data with date-bounded accuracy.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_historical_zero_dteGet Historical Zero-DTEA
Read-only
Inspect

Replay 0DTE analytics (pin risk, expected move, gamma acceleration, dealer hedging estimates for same-day expiry) at any minute since April 2018. Alpha tier.

ParametersJSON Schema
NameRequiredDescriptionDefault
atYesAs-of timestamp: YYYY-MM-DDTHH:mm:ss (ET) or YYYY-MM-DD
apiKeyYesYour FlashAlpha API key (Alpha tier)
symbolYesStock/ETF ticker
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations provide readOnlyHint=true, and the description adds valuable behavioral context beyond this: it specifies the data availability timeframe ('since April 2018'), indicates it's for 'same-day expiry' analytics, and mentions the 'Alpha tier' requirement. No contradiction with annotations exists.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is efficiently structured in a single sentence that packs essential information: action, resource, scope, analytics list, timeframe, and tier requirement. Every element serves a clear purpose with zero wasted words.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a read-only tool with full parameter documentation but no output schema, the description provides good context about what analytics are returned and historical scope. However, it doesn't describe the return format or structure, leaving some ambiguity about the output.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

With 100% schema description coverage, the input schema already fully documents all three parameters. The description doesn't add any parameter-specific information beyond what's in the schema, so it meets the baseline of 3 for high schema coverage.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the specific action ('Replay 0DTE analytics') and resource ('at any minute since April 2018'), listing the exact analytics provided (pin risk, expected move, gamma acceleration, dealer hedging estimates). It distinguishes from siblings like 'get_zero_dte' by specifying historical data and time range.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides clear context for when to use this tool ('Replay 0DTE analytics... at any minute since April 2018'), implying it's for historical data rather than current. However, it doesn't explicitly state when NOT to use it or name specific alternatives among the many sibling tools.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_levelsGet Key Options LevelsB
Read-only
Inspect

Get key options levels: gamma flip point, call wall, put wall, max pain, highest OI strike. These act as support/resistance from dealer hedging.

ParametersJSON Schema
NameRequiredDescriptionDefault
apiKeyYesYour FlashAlpha API key
symbolYesStock/ETF ticker
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations indicate readOnlyHint=true, which the description aligns with by using 'Get' (implying a read operation). The description adds context by explaining that the retrieved levels 'act as support/resistance from dealer hedging', which provides behavioral insight beyond the annotations. However, it doesn't disclose other traits like rate limits, error conditions, or data freshness, leaving some gaps in transparency.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, efficient sentence that lists the key options levels and their functional context. It's front-loaded with the core purpose and avoids unnecessary details. While very concise, it could potentially benefit from slightly more structure (e.g., separating the list for clarity), but it earns its place without waste.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's moderate complexity (retrieving calculated options levels), the description covers the purpose and adds some behavioral context. However, with no output schema and annotations only providing readOnlyHint, it lacks details on return format, data types, or example outputs. The description doesn't fully compensate for these gaps, making it adequate but incomplete for optimal agent use.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The input schema has 100% description coverage, with clear documentation for both parameters (apiKey and symbol). The description doesn't add any parameter-specific semantics beyond what the schema provides, such as format examples or constraints. According to the rules, with high schema coverage (>80%), the baseline score is 3, which is appropriate here as the description doesn't compensate but doesn't need to.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool retrieves specific options levels (gamma flip point, call wall, put wall, max pain, highest OI strike) and mentions their function as support/resistance from dealer hedging. It uses the verb 'Get' with the resource 'key options levels', making the purpose explicit. However, it doesn't explicitly differentiate from sibling tools like get_option_chain or get_advanced_volatility, which might also provide options-related data.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides no guidance on when to use this tool versus alternatives. It doesn't mention any prerequisites, exclusions, or specific contexts for use. Given the many sibling tools related to options and volatility (e.g., get_option_chain, get_advanced_volatility, get_gex), the lack of differentiation leaves the agent without clear usage instructions.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_narrativeGet GEX NarrativeA
Read-only
Inspect

Get verbal GEX narrative analysis. Describes gamma regime, key levels, dealer positioning, and price action implications in plain English.

ParametersJSON Schema
NameRequiredDescriptionDefault
apiKeyYesYour FlashAlpha API key
symbolYesStock/ETF ticker
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations provide readOnlyHint=true, indicating a safe read operation. The description adds context about the output format (plain English narrative analysis covering specific aspects), which is valuable beyond annotations. However, it doesn't mention rate limits, authentication needs beyond the apiKey parameter, or error behavior.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, well-structured sentence that efficiently conveys the tool's purpose and output characteristics. Every word adds value, with no redundant or vague phrasing.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's moderate complexity (financial analysis output), the description adequately covers what the tool does and its output format. However, without an output schema, it could benefit from more detail on the narrative structure or example output. Annotations help by indicating read-only behavior.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, with clear descriptions for both parameters (apiKey and symbol). The description doesn't add any additional parameter semantics beyond what the schema provides, such as format examples or constraints, so the baseline score of 3 is appropriate.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the verb 'Get' and resource 'verbal GEX narrative analysis', specifying it describes gamma regime, key levels, dealer positioning, and price action implications in plain English. This distinguishes it from sibling tools like 'get_gex' (likely raw data) and 'get_exposure_summary' (different focus).

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage when a verbal analysis of GEX is needed, but doesn't explicitly state when to use this tool versus alternatives like 'get_gex' (which might provide raw data) or other analysis tools. No exclusions or prerequisites are mentioned.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_option_chainGet Option ChainA
Read-only
Inspect

Get option chain metadata: available expirations and strikes for a ticker.

ParametersJSON Schema
NameRequiredDescriptionDefault
apiKeyYesYour FlashAlpha API key
symbolYesStock/ETF ticker
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations declare readOnlyHint=true, indicating a safe read operation. The description adds context by specifying the type of metadata returned (expirations and strikes), which is useful beyond the annotations. However, it does not disclose behavioral traits like rate limits, authentication needs (implied by apiKey but not stated), or response format, leaving gaps in transparency.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, efficient sentence that front-loads the core purpose ('Get option chain metadata') and specifies the scope ('available expirations and strikes for a ticker'). There is no wasted verbiage, making it highly concise and well-structured.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's moderate complexity (2 parameters, no output schema), the description is minimally complete. It states what the tool does but lacks details on output format, error handling, or integration with siblings. With annotations covering safety, it meets basic needs but leaves room for improvement in guiding effective use.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, with clear descriptions for both parameters (apiKey and symbol). The description adds no additional meaning beyond the schema, such as format details or usage examples. Baseline score of 3 is appropriate as the schema adequately documents parameters without extra value from the description.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the specific action ('Get') and resource ('option chain metadata'), specifying what metadata is retrieved ('available expirations and strikes for a ticker'). It distinguishes from siblings like 'get_option_quote' (which retrieves quotes rather than chain metadata) and 'get_historical_option_quote' (historical data).

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides no guidance on when to use this tool versus alternatives. It does not mention when to choose this over 'get_option_quote' (for specific option data) or 'calculate_greeks' (for derived metrics), nor does it specify prerequisites or exclusions, leaving usage context implied at best.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_option_quoteGet Option QuoteA
Read-only
Inspect

Get live option quote with bid, ask, mid, IV, greeks, open interest, and volume. Filter by expiry, strike, and type.

ParametersJSON Schema
NameRequiredDescriptionDefault
typeNo'C' or 'P' (call or put)
apiKeyYesYour FlashAlpha API key
expiryNoExpiration date YYYY-MM-DD
strikeNoStrike price
symbolYesUnderlying ticker (e.g. SPY, AAPL)
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations provide readOnlyHint=true, which the description aligns with by describing a data retrieval operation. The description adds valuable behavioral context beyond annotations by specifying it returns 'live' data (real-time/current), listing the specific fields returned (bid, ask, etc.), and mentioning filtering capabilities. It doesn't contradict annotations, but could mention rate limits or authentication requirements more explicitly.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, well-structured sentence that front-loads the core purpose and efficiently lists both the returned data fields and filter parameters. Every element serves a clear purpose with zero wasted words, making it easy to parse quickly.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a read-only tool with good annotations and full schema coverage, the description provides adequate context by specifying it's for live quotes and listing return fields. However, without an output schema, it could more explicitly describe the response structure or data formats. The sibling context is partially addressed through implicit differentiation.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, so all parameters are documented in the schema. The description adds minimal value beyond the schema by mentioning the three filter parameters (expiry, strike, type) but doesn't provide additional semantic context like format examples for 'symbol' or explain parameter interactions. Baseline 3 is appropriate given the comprehensive schema coverage.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the action ('Get live option quote') and specifies the exact data returned (bid, ask, mid, IV, greeks, open interest, volume). It distinguishes from siblings like get_option_chain (which returns multiple options) and get_historical_option_quote (which returns historical data) by focusing on a single live quote with filtering capabilities.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage context through the filter parameters (expiry, strike, type), suggesting this is for retrieving specific option quotes. However, it doesn't explicitly state when to use this versus alternatives like get_option_chain (for multiple options) or get_historical_option_quote (for historical data), nor does it mention prerequisites like needing an API key.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_stock_quoteGet Stock QuoteA
Read-only
Inspect

Get real-time stock quote (bid, ask, mid, last price) for a ticker symbol.

ParametersJSON Schema
NameRequiredDescriptionDefault
apiKeyYesYour FlashAlpha API key
symbolYesStock ticker (e.g. SPY, AAPL, TSLA)
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations provide readOnlyHint=true, indicating a safe read operation. The description adds value by specifying the real-time nature and the specific quote fields returned (bid, ask, mid, last price), which are not covered by annotations. However, it doesn't disclose behavioral aspects like rate limits, authentication needs beyond the apiKey parameter, or error handling.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, efficient sentence that front-loads the core purpose ('Get real-time stock quote') and includes essential details (data fields and resource) without any wasted words. Every part of the sentence contributes directly to understanding the tool's function.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's low complexity (2 parameters, no output schema), the description is reasonably complete. It covers the purpose, real-time aspect, and returned data fields. However, without an output schema, it could benefit from more detail on the response structure, but the annotations and schema provide adequate support for basic use.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, with clear descriptions for both parameters (apiKey and symbol). The description adds no additional parameter semantics beyond what the schema provides, such as format details or usage examples. With high schema coverage, the baseline score of 3 is appropriate as the description doesn't compensate but also doesn't detract.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the specific action ('Get real-time stock quote') and resource ('for a ticker symbol'), with explicit mention of the data fields returned (bid, ask, mid, last price). It distinguishes from siblings like get_historical_stock_quote by specifying 'real-time' and from get_stock_summary by focusing on quote data rather than summary information.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage context through 'real-time' and 'ticker symbol', suggesting when to use this tool versus historical alternatives. However, it lacks explicit guidance on when not to use it or direct alternatives among the many sibling tools, leaving some ambiguity for the agent.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_stock_summaryGet Stock SummaryB
Read-only
Inspect

Get comprehensive stock summary: price, ATM IV, historical vol, VRP, skew, term structure, options flow, exposure data, and macro context (VIX, Fear & Greed, yield curve).

ParametersJSON Schema
NameRequiredDescriptionDefault
apiKeyYesYour FlashAlpha API key
symbolYesStock/ETF/index ticker (e.g. SPY, AAPL, SPX)
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations provide readOnlyHint=true, indicating a safe read operation. The description adds value by specifying the scope of data returned (e.g., price, VRP, macro context), which isn't covered by annotations. However, it doesn't disclose behavioral traits like rate limits, error handling, or data freshness, leaving gaps despite the annotation coverage.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, efficient sentence that front-loads the key action ('Get comprehensive stock summary') and lists data components without unnecessary words. It could be slightly more structured by grouping related data points, but it's appropriately sized and avoids redundancy.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's complexity (returning multiple data types) and lack of output schema, the description provides a good overview of what data to expect. However, it doesn't specify format, units, or how data is organized, which could be crucial for an AI agent. With annotations covering safety, it's adequate but not fully complete.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, with clear documentation for both parameters (symbol and apiKey). The description doesn't add any parameter-specific details beyond what the schema provides, such as format examples or constraints. Baseline 3 is appropriate since the schema handles the heavy lifting.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool's purpose with a specific verb ('Get') and resource ('comprehensive stock summary'), listing detailed data components like price, ATM IV, VRP, and macro context. It distinguishes itself from simpler siblings like 'get_stock_quote' by emphasizing comprehensiveness, though it doesn't explicitly name alternatives.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage for comprehensive stock analysis by listing extensive data points, suggesting it's for detailed summaries rather than basic quotes. However, it lacks explicit guidance on when to use this vs. alternatives like 'get_stock_quote' or 'get_exposure_summary', and doesn't mention prerequisites or exclusions.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_tickersList Available TickersB
Read-only
Inspect

List all available stock/ETF tickers with live options data.

ParametersJSON Schema
NameRequiredDescriptionDefault
apiKeyYesYour FlashAlpha API key
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations already declare readOnlyHint=true, indicating a safe read operation. The description adds context by specifying 'live options data', which hints at real-time or current data, but doesn't disclose behavioral traits like rate limits, pagination, or data freshness. With annotations covering safety, it adds some value but not rich behavioral details.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, efficient sentence that directly states the tool's purpose without unnecessary words. It's front-loaded and every part earns its place, making it highly concise and well-structured.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's low complexity (one parameter, read-only), no output schema, and good annotations, the description is adequate but incomplete. It doesn't cover aspects like return format (e.g., list structure, data fields) or potential limitations, which could help the agent use it more effectively.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, with the single parameter 'apiKey' fully documented in the schema. The description doesn't add any meaning beyond the schema, such as explaining why the API key is needed or how it's used, so it meets the baseline for high coverage.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the verb 'List' and the resource 'stock/ETF tickers with live options data', making the purpose specific and understandable. However, it doesn't explicitly differentiate from sibling tools like 'get_stock_summary' or 'get_option_chain', which might also involve ticker information, so it doesn't reach the highest score.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides no guidance on when to use this tool versus alternatives. It doesn't mention prerequisites like needing an API key or compare it to siblings such as 'get_stock_quote' or 'get_option_chain', leaving the agent to infer usage context.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_vexGet Vanna Exposure (VEX)B
Read-only
Inspect

Get vanna exposure (VEX) by strike. Shows how dealer hedging changes with volatility moves.

ParametersJSON Schema
NameRequiredDescriptionDefault
apiKeyYesYour FlashAlpha API key
symbolYesStock/ETF ticker
expirationNoOptional expiration date YYYY-MM-DD
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations provide readOnlyHint=true, indicating a safe read operation. The description adds context about what VEX measures ('how dealer hedging changes with volatility moves'), which is valuable beyond the annotations. However, it doesn't disclose behavioral traits like rate limits, authentication needs beyond the apiKey parameter, or what happens with invalid inputs. No contradiction with annotations exists.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is extremely concise with two clear sentences that front-load the core purpose. Every word earns its place, with no redundant information. The structure moves from the basic function to the specific insight provided, making it easy to scan and understand quickly.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a read-only tool with good annotations and full schema coverage, the description provides adequate basic context. However, without an output schema, it doesn't explain what the tool returns (e.g., data format, structure, or sample values). Given the specialized financial nature of VEX and many sibling alternatives, more context about the output and use cases would be beneficial.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, so the schema fully documents all parameters (symbol, apiKey, expiration). The description adds no additional parameter semantics beyond what's in the schema. It mentions 'by strike' which might imply strike-related filtering, but this isn't reflected in the parameters, creating potential confusion rather than clarity.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool's purpose: 'Get vanna exposure (VEX) by strike' with the specific action 'shows how dealer hedging changes with volatility moves.' It distinguishes from siblings by focusing on VEX rather than other metrics like Greeks, volatility, or quotes. However, it doesn't explicitly contrast with similar exposure tools like get_gex or get_exposure_summary.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides no guidance on when to use this tool versus alternatives. It mentions 'by strike' but doesn't clarify if this is for single strikes, ranges, or all strikes. With many sibling tools for calculations, exposures, and quotes, there's no indication of when VEX is preferred over other exposure metrics or calculation tools.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_volatilityGet Volatility AnalysisA
Read-only
Inspect

Get comprehensive volatility analysis: ATM IV, realized vol (5/10/20/30d), VRP, 25-delta skew, IV term structure, GEX by DTE, theta by DTE, hedging scenarios, liquidity metrics.

ParametersJSON Schema
NameRequiredDescriptionDefault
apiKeyYesYour FlashAlpha API key
symbolYesStock/ETF ticker
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations provide readOnlyHint=true, indicating a safe read operation. The description adds valuable context by specifying the comprehensive set of metrics returned (e.g., ATM IV, realized vol, VRP, skew), which goes beyond annotations to clarify the tool's behavioral output. No contradictions with annotations.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, dense sentence that front-loads the purpose ('Get comprehensive volatility analysis') and efficiently lists all key metrics without unnecessary words. Every part earns its place by informing the user of the tool's scope.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's complexity (multiple volatility metrics), annotations cover safety (read-only), and schema fully documents parameters, the description provides a clear overview of what analyses are included. However, without an output schema, it could benefit from more detail on return format or data structure to be fully complete.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, with clear descriptions for both parameters (apiKey and symbol). The description doesn't add parameter-specific semantics beyond what the schema provides, such as symbol format examples or apiKey usage details, meeting the baseline for high schema coverage.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description explicitly states 'Get comprehensive volatility analysis' followed by a detailed list of specific metrics (ATM IV, realized vol, VRP, skew, etc.), clearly indicating it retrieves multiple volatility-related analyses for a given symbol. This distinguishes it from sibling tools like get_vrp (single metric) or get_advanced_volatility (likely more complex).

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage for volatility analysis needs but doesn't explicitly state when to use this tool versus alternatives like get_vrp (for VRP only) or get_advanced_volatility. It provides context through the listed metrics but lacks explicit guidance on tool selection or exclusions.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_vrpGet VRP DashboardA
Read-only
Inspect

Get volatility risk premium (VRP) dashboard: live IV vs realized vol, VRP percentiles, term structure, regime classification, strategy scores, and macro context.

ParametersJSON Schema
NameRequiredDescriptionDefault
apiKeyYesYour FlashAlpha API key
symbolYesStock/ETF ticker
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations indicate readOnlyHint=true, which the description aligns with by using 'Get' (a read operation). The description adds value beyond annotations by detailing the specific components of the dashboard (e.g., live IV vs realized vol, VRP percentiles), which helps the agent understand the behavioral output. However, it does not disclose additional traits like rate limits or error handling.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, efficient sentence that front-loads the purpose and lists dashboard components without unnecessary words. Every part of the sentence contributes to understanding the tool's functionality, making it highly concise and well-structured.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's complexity (providing a detailed dashboard) and lack of output schema, the description does a good job of outlining what the dashboard includes. However, it could be more complete by mentioning the format of the return data or any limitations. The annotations cover the read-only aspect, but additional context on output structure would enhance completeness.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, with clear descriptions for both parameters (apiKey and symbol). The description does not add meaning beyond the schema, as it focuses on the dashboard content rather than parameter usage. Baseline score is 3 since the schema adequately covers parameter semantics.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the verb 'Get' and the resource 'volatility risk premium (VRP) dashboard', with specific details about what the dashboard includes: live IV vs realized vol, VRP percentiles, term structure, regime classification, strategy scores, and macro context. It distinguishes from sibling tools like 'get_vrp_history' by focusing on a current dashboard rather than historical data.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage for obtaining a comprehensive VRP dashboard, but it does not explicitly state when to use this tool versus alternatives like 'get_volatility' or 'get_vrp_history'. It provides context about the dashboard content but lacks explicit guidance on exclusions or prerequisites beyond the required parameters.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_vrp_historyGet VRP HistoryA
Read-only
Inspect

Get historical VRP time series: daily ATM IV, realized vol (5/10/20/30d), VRP, straddle price, and expected move for charting and backtesting.

ParametersJSON Schema
NameRequiredDescriptionDefault
daysNoNumber of days of history (default 30, max 365)
apiKeyYesYour FlashAlpha API key
symbolYesStock/ETF ticker
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations already declare readOnlyHint=true, indicating a safe read operation. The description adds valuable context by specifying the data types returned (time series with specific metrics) and the purpose (charting and backtesting), which helps the agent understand the tool's behavior beyond the annotations, though it lacks details on rate limits or response format.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, efficient sentence that front-loads the core purpose and lists the returned data points without unnecessary words, making it easy to parse and understand quickly.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's complexity (historical data retrieval with multiple metrics) and the absence of an output schema, the description adequately covers what data is returned and its purpose. However, it could be more complete by mentioning the response structure or any limitations, such as data availability or format.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, so the schema fully documents the parameters (symbol, apiKey, days). The description does not add any additional meaning or clarification about the parameters beyond what the schema provides, such as examples or constraints, meeting the baseline for high schema coverage.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the specific action ('Get historical VRP time series') and enumerates the exact data points returned (daily ATM IV, realized vol for multiple periods, VRP, straddle price, expected move), distinguishing it from sibling tools like 'get_vrp' or 'get_volatility' by emphasizing historical data for charting and backtesting purposes.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage for 'charting and backtesting,' which provides some context, but it does not explicitly state when to use this tool versus alternatives like 'get_vrp' or 'get_historical_stock_quote,' nor does it mention any exclusions or prerequisites beyond the required parameters.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_zero_dteGet Zero-DTE AnalyticsA
Read-only
Inspect

Get zero-days-to-expiration (0DTE) analytics: intraday gamma, time decay acceleration, pin risk, dealer hedging pressure for contracts expiring today.

ParametersJSON Schema
NameRequiredDescriptionDefault
apiKeyYesYour FlashAlpha API key
symbolYesStock/ETF ticker
strike_rangeNoStrike range as decimal fraction of spot (default 0.03 = 3%)
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations provide readOnlyHint=true, indicating a safe read operation. The description adds valuable context beyond annotations by specifying the analytics types returned (gamma, time decay acceleration, etc.) and the expiration constraint ('contracts expiring today'). It doesn't mention rate limits, authentication requirements (though apiKey is in schema), or response format details, but provides meaningful behavioral information about what data to expect.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, efficient sentence that front-loads the purpose and key details. Every word earns its place: it defines the acronym (0DTE), lists the specific analytics returned, and specifies the expiration constraint. There's no wasted verbiage or redundancy with the title or schema.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a read-only tool with good annotations and 100% schema coverage, the description provides adequate context about what analytics are returned and their scope. The main gap is the lack of output schema, so the description doesn't explain the structure or format of returned data. However, given the tool's relatively straightforward purpose and good parameter documentation, the description is reasonably complete for agent selection.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, so the schema already documents all parameters thoroughly. The description doesn't add any parameter-specific information beyond what's in the schema (e.g., it doesn't explain symbol format or strike_range implications). With complete schema coverage, the baseline score of 3 is appropriate as the description doesn't enhance parameter understanding but doesn't need to compensate for gaps.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the specific action ('Get') and resource ('zero-days-to-expiration (0DTE) analytics') with precise scope ('for contracts expiring today'). It lists specific analytics types (intraday gamma, time decay acceleration, pin risk, dealer hedging pressure) that distinguish it from sibling tools like 'get_advanced_volatility' or 'get_volatility' which likely focus on different volatility metrics.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage context through 'for contracts expiring today' and the analytics types mentioned, suggesting this is for intraday analysis of near-expiration options. However, it doesn't explicitly state when to use this tool versus alternatives like 'get_option_chain' or 'get_advanced_volatility', nor does it mention any prerequisites or exclusions beyond the expiration timeframe.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

solve_ivSolve Implied VolatilityA
Read-only
Inspect

Solve for implied volatility from option market price. Reverse-engineers BSM to find what vol is priced in.

ParametersJSON Schema
NameRequiredDescriptionDefault
dteYesDays to expiration
spotYesCurrent stock price
typeYes'call' or 'put'
priceYesOption market price
apiKeyYesYour FlashAlpha API key
strikeYesStrike price
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations indicate readOnlyHint=true, confirming this is a safe read operation. The description adds behavioral context by specifying it 'reverse-engineers BSM to find what vol is priced in,' which clarifies the computational method. However, it doesn't disclose rate limits, error handling, or output format, leaving gaps in behavioral understanding.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is front-loaded and highly concise, consisting of two efficient sentences that directly state the tool's purpose and method. Every word contributes value, with no wasted information, making it easy for an agent to parse quickly.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's complexity (mathematical calculation with 6 parameters) and lack of output schema, the description is minimally adequate. It explains the core function but omits details on return values, error conditions, or integration with sibling tools, leaving room for improvement in completeness.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, with all parameters clearly documented in the input schema. The description adds no additional parameter semantics beyond implying the use of BSM model inputs, so it meets the baseline of 3 without compensating for any gaps.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the specific action ('solve for implied volatility') and resource ('from option market price'), distinguishing it from siblings like 'calculate_greeks' or 'get_volatility' by focusing on reverse-engineering the Black-Scholes-Merton (BSM) model. It precisely defines the tool's mathematical purpose without redundancy.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides no guidance on when to use this tool versus alternatives, such as 'get_volatility' or 'get_advanced_volatility', nor does it mention prerequisites like needing an API key or valid option data. Usage is implied only through the mathematical context, lacking explicit when/when-not instructions.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

Discussions

No comments yet. Be the first to start the discussion!

Try in Browser

Your Connectors

Sign in to create a connector for this server.