Skip to main content
Glama

Agent Einstein — Autonomous Crypto Intelligence

Server Details

40 AI crypto tools: whale tracking, security scans, DeFi analytics, quantum security, and more.

Status
Healthy
Last Tested
Transport
Streamable HTTP
URL

Glama MCP Gateway

Connect through Glama MCP Gateway for full control over tool access and complete visibility into every call.

MCP client
Glama
MCP server

Full call logging

Every tool call is logged with complete inputs and outputs, so you can debug issues and audit what your agents are doing.

Tool access control

Enable or disable individual tools per connector, so you decide what your agents can and cannot do.

Managed credentials

Glama handles OAuth flows, token storage, and automatic rotation, so credentials never expire on your clients.

Usage analytics

See which tools your agents call, how often, and when, so you can understand usage patterns and catch anomalies.

100% free. Your data is private.

Tool Definition Quality

Score is being calculated. Check back soon.

Available Tools

47 tools
arbitrage_scannerCInspect

Arbitrage Scanner: Cross-chain and cross-DEX price discrepancies for arbitrage opportunities.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description carries the full burden of behavioral disclosure. It fails to indicate whether the tool executes trades or merely scans, what the output format contains, data latency/freshness, or financial risk considerations. It also does not disclose whether the operation is read-only (implied by 'scanner' but not confirmed) or requires specific permissions.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, efficient sentence that front-loads the core concept. The 'Arbitrage Scanner:' prefix is slightly redundant with the tool name but does not significantly detract from clarity. No extraneous information is included.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given this is a financial analysis tool with no output schema and no annotations, the description is incomplete. It fails to describe the return value (e.g., opportunity details, profit calculations, risk metrics), does not mention execution versus read-only behavior, and omits safety considerations typical for arbitrage operations.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The description does not explicitly mention any parameters, but the input schema has 100% description coverage with clear enum values for 'chain' and 'timeperiod'. Since the schema fully documents all three parameters, the baseline score applies; the description adds no additional semantic context beyond what the schema provides.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly identifies the tool's function: scanning for price discrepancies across chains and DEXes to identify arbitrage opportunities. It uses specific verbs and resources ('scan', 'price discrepancies', 'arbitrage opportunities'). However, it fails to distinguish itself from the sibling tool 'cross_chain_arbitrage', which creates potential selection ambiguity.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides no guidance on when to use this tool versus alternatives, particularly the similarly-named 'cross_chain_arbitrage' sibling. It does not mention prerequisites (e.g., requiring a specific token address), nor does it indicate whether this tool is for analysis only or execution.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

bitcoin_bridge_comparisonBInspect

Bitcoin Bridge Comparison: Compare Bitcoin bridges by volume, track wrapped BTC peg deviations (wBTC, cbBTC, tBTC), and view THORChain BTC pool stats.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full burden. It discloses the data domains covered (volume metrics, peg deviations, pool stats) but lacks operational details such as data freshness, rate limiting, error handling, or whether this triggers any on-chain transactions. It implies read-only behavior but doesn't state it explicitly.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single efficient sentence that front-loads the subject matter. The colon-separated structure ('Bitcoin Bridge Comparison: Compare...') is slightly redundant with the tool name, but every clause adds specific detail about data types (volume, peg deviations, pool stats) without waste.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given that all three parameters are optional and no output schema exists, the description adequately identifies the subject matter but lacks critical operational context. It should specify default behavior when called without parameters (since required: []) and ideally hint at the return structure (e.g., 'returns comparative metrics including...').

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, establishing a baseline of 3. The description mentions 'volume' and time-based tracking which loosely maps to the 'timeperiod' parameter, and specific chains map to 'chain', but it doesn't explicitly connect these concepts or explain parameter interactions (e.g., that 'chain' filters the bridge comparison to that specific network).

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly specifies the verb (compare, track, view) and resources (Bitcoin bridges, wBTC/cbBTC/tBTC pegs, THORChain pools). It provides concrete asset examples that distinguish it from generic analytics tools. However, it doesn't explicitly contrast usage with siblings like 'cross_chain_arbitrage' or 'bitcoin_onchain_analytics'.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance is provided on when to use this tool versus alternatives, prerequisites for use, or when to avoid it. The description states what data is available but not the analytical scenarios where this tool is appropriate (e.g., 'use this to assess bridge risk' or 'use arbitrage_scanner instead for pricing discrepancies').

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

bitcoin_mempool_intelligenceCInspect

Bitcoin Mempool & Fee Intelligence: Real-time Bitcoin mempool analytics, fee estimation, and transaction tracking. Includes congestion level, recommended fee tiers, and confirmation ETA.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Without annotations, the description carries the full burden. It partially compensates by disclosing what data is returned (congestion levels, fee tiers, confirmation ETA), which is helpful given the lack of an output schema. However, it omits operational details like rate limits, authentication requirements, or caching behavior.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is appropriately concise with two front-loaded sentences. The first establishes the domain and the second lists specific capabilities. No sentences are wasted, though the first sentence functions somewhat as a title重复 since the tool name is already descriptive.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the lack of output schema, the description adequately explains what data is returned (congestion, fees, ETA). However, the Bitcoin-specific claim creates confusion given the multi-chain schema, and the description omits whether the data is real-time vs historical despite mentioning 'real-time' in the text.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters2/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

While the schema has 100% coverage (baseline 3), the description detracts value by implying Bitcoin-only usage when the 'chain' parameter explicitly supports Ethereum, Base, Solana, and others. It adds no clarifying context for the 'limit' or 'timeperiod' parameters beyond what the schema provides.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose3/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool provides Bitcoin mempool analytics, fee estimation, and transaction tracking with specific outputs (congestion level, fee tiers, ETA), which distinguishes it from siblings like bitcoin_onchain_analytics. However, it claims Bitcoin-specific functionality while the input schema's 'chain' parameter accepts Ethereum, Solana, and other non-Bitcoin networks, creating a significant scope ambiguity.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides no guidance on when to use this tool versus alternatives like bitcoin_onchain_analytics or arbitrage_scanner. It does not mention prerequisites, such as whether this requires a specific Bitcoin node connection, or when not to use it.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

bitcoin_onchain_analyticsCInspect

Bitcoin Network & Mining Analytics: Mining pool distribution, hashrate trends, difficulty adjustment tracking, block analytics, and Bitcoin address lookup.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations provided, so description carries full burden. It lists data categories returned but omits critical behavioral details: whether data is real-time or historical, latency expectations, what happens when no parameters are provided (all are optional), or whether the tool is read-only.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness3/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Single sentence is appropriately concise and front-loaded with the core topic, but includes unsupported claims (address lookup) that reduce the value of the limited space used.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Incomplete for a tool with 3 parameters and rich siblings. Fails to resolve the Bitcoin/multi-chain schema mismatch, explain the missing address parameter for claimed lookup functionality, or describe output format (no output schema provided).

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters2/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

While schema coverage is 100% (baseline 3), the description claims 'address lookup' functionality but no address parameter exists in the schema. It also fails to explain why a 'bitcoin' tool accepts non-Bitcoin chains (base, ethereum, solana) in the chain parameter, creating semantic confusion.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose3/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly lists Bitcoin-specific capabilities (mining pools, hashrate, difficulty tracking), but fails to address the contradiction between the Bitcoin-specific name/description and the multi-chain schema (which includes Ethereum, Solana, etc. in the chain parameter enum). It also doesn't distinguish from sibling tools like bitcoin_mempool_intelligence or bitcoin_staking_analysis.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance provided on when to select this tool versus sibling alternatives (e.g., bitcoin_mempool_intelligence for mempool data, general_analysis for broader crypto). No prerequisites mentioned despite all parameters being optional.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

bitcoin_staking_analysisBInspect

Babylon Bitcoin Staking Analytics: Babylon staking analytics including global TVL stats, finality provider leaderboards, staker delegation positions, and individual delegation tracking.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full burden. It discloses what data is returned (global TVL, leaderboards, positions), but omits operational details like authentication requirements, rate limits, cache behavior, or whether the analysis is real-time versus historical.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Single sentence efficiently lists the four data categories provided. Minor redundancy in repeating 'Babylon ... Analytics' and 'Babylon ... analytics' with the colon structure, but otherwise no wasted words.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Adequate for an analytics tool with no output schema—compensates by enumerating the return data types (TVL, leaderboards, etc.). However, given the complex domain (Bitcoin staking with Babylon protocol) and lack of annotations, it would benefit from explicit usage context or behavioral constraints.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The input schema has 100% description coverage for all three parameters (chain, limit, timeperiod). The description adds no additional parameter context, but since the schema is self-documenting, this meets the baseline expectation.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

Clearly identifies the tool provides Babylon Bitcoin staking analytics (TVL stats, finality provider leaderboards, delegation tracking), distinguishing it from siblings like bitcoin_onchain_analytics and brc20_analytics. However, it uses a noun phrase structure ('Analytics: ... analytics') rather than a specific action verb, slightly weakening clarity.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Provides no guidance on when to use this tool versus the numerous sibling alternatives (bitcoin_onchain_analytics, runes_analytics, etc.) or what prerequisites might exist. No mention of when not to use it or complementary tools.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

brc20_analyticsCInspect

BRC-20 Token Analytics: BRC-20 token overview, mint progress, holder distribution, and individual token details via UniSat API.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions 'via UniSat API' indicating the data source, but fails to disclose whether the tool is read-only or destructive, rate limits, authentication requirements, or error behaviors. For an analytics tool with no safety annotations, this is insufficient behavioral coverage.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single efficient sentence with no redundant words. The key resource (BRC-20) is front-loaded. However, it could benefit from structured formatting (e.g., separating capability list from API source) to improve scannability for agents parsing the text.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the absence of an output schema, the description compensates partially by listing the types of data returned (overview, mint progress, holder distribution). However, with three optional parameters and no annotations, it should clarify default behaviors when parameters are omitted or explain the UniSat API constraints. It provides the minimum necessary context but leaves operational gaps.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

With 100% schema description coverage, the baseline score applies. The description does not mention any parameters or provide context for why one might filter by 'chain' (particularly notable since BRC-20 is Bitcoin-native yet the schema allows Ethereum/Solana chains), 'limit', or 'timeperiod'. However, since the schema fully documents all three parameters, the description meets the minimum viable threshold.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly identifies the tool's function as BRC-20 token analytics and lists specific data points provided (overview, mint progress, holder distribution, individual token details). It distinguishes from sibling tools like 'runes_analytics' by explicitly naming the BRC-20 standard and UniSat API data source. However, it lacks explicit comparative guidance against the similar Runes analytics sibling.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides no guidance on when to use this tool versus alternatives (e.g., 'runes_analytics' for the Runes token standard), nor does it mention prerequisites, required conditions, or when not to use it. Agents must infer applicability solely from the resource name.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

copy_trading_intelCInspect

Copy Trading Intelligence: Analyze top-performing wallets for copy trading signals with risk-adjusted position sizing.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full burden of behavioral disclosure. It mentions 'risk-adjusted position sizing' hinting at analytical complexity, but fails to clarify whether this is a read-only operation (likely), what specific data fields are returned, how risk is calculated, or any rate limiting concerns. This leaves significant behavioral gaps for a financial analysis tool.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, efficient sentence that avoids verbosity. The 'Copy Trading Intelligence:' prefix functions as a conceptual header that slightly weakens the front-loading, but the core content immediately follows. No redundant or filler text is present.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the absence of annotations and output schema, the description should explain what the analysis returns (e.g., wallet rankings, performance metrics, position size recommendations). It also fails to clarify the relationship with 'copy_trading_leaderboard'. For a complex financial tool with zero structured metadata, the description is insufficiently complete.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The description does not explicitly mention any input parameters (chain, limit, timeperiod). However, the input schema has 100% description coverage with clear enums and types, establishing a baseline of 3. Since the schema fully documents the parameters, the description does not need to compensate, though it adds no additional semantic context about parameter interactions.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool analyzes top-performing wallets to generate copy trading signals with risk-adjusted position sizing. It specifies the verb (analyze), resource (wallets), and output type (signals with sizing). However, it does not explicitly differentiate from the sibling tool 'copy_trading_leaderboard', leaving ambiguity about when to choose this over the alternative.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides no guidance on when to use this tool versus alternatives, prerequisites for use, or conditions where it should be avoided. Given the presence of 'copy_trading_leaderboard' and 'smart_money' siblings, the lack of explicit when-to-use guidance forces the agent to guess based on naming alone.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

copy_trading_leaderboardBInspect

Copy Trading Leaderboard: Top wallets worth copying with performance data, win rates, and P&L metrics.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description carries the full disclosure burden. It compensates partially by listing return data types (performance metrics, win rates, P&L) since no output schema exists. However, it lacks operational details such as ranking methodology, data freshness, rate limits, or whether results are cached.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, efficient sentence with zero wasted words. The colon structure front-loads the concept (Leaderboard) and follows with specific data attributes, making it appropriately sized for the tool's complexity.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the absence of an output schema, the description adequately compensates by specifying the categories of returned metrics. However, with zero required parameters and three optional filters, it could better explain the default behavior when no parameters are provided or how the ranking algorithm weighs different performance factors.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

With 100% schema description coverage, the baseline score is 3. The description does not add semantic context to the parameters (e.g., explaining how `timeperiod` affects metric calculations or that `chain` filters the universe of wallets), but it does not need to given the comprehensive schema documentation.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly identifies the resource (Copy Trading Leaderboard) and specific data returned (performance data, win rates, P&L metrics). However, it does not differentiate from the sibling tool `copy_trading_intel` or clarify when to prefer this leaderboard view over other trading analysis tools.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides no guidance on when to use this tool versus alternatives like `copy_trading_intel`, `smart_money`, or `wallet_trust_score`. It does not mention prerequisites, filtering strategies, or limitations of the leaderboard approach.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

cross_chain_arbitrageAInspect

Cross-Chain Arbitrage Scanner: Detect price discrepancies across chains and DEXes with net profit calculations after bridge fees.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full burden of behavioral disclosure. It successfully communicates that the tool performs calculations (net profit after bridge fees) rather than just raw price comparison, and 'detect' implies a read-only operation. However, it omits operational details like data freshness, rate limits, or authentication requirements that would typically appear in annotations.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, efficiently structured sentence. The prefix 'Cross-Chain Arbitrage Scanner:' acts as an immediate identifier, followed by the specific capability. Every clause adds value, with zero redundancy or filler content.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a 3-parameter tool with no required fields and simple types, the description adequately covers the functional scope. It compensates for the lack of output schema by indicating what calculations are returned (net profit). Minor gaps remain regarding operational constraints, but the description is sufficient for tool selection given the moderate complexity.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The input schema has 100% description coverage with clear enum values and types. The description does not explicitly discuss individual parameters, but with complete schema documentation, no additional parameter semantics are necessary. Baseline score applies.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the specific action (detect price discrepancies), the scope (across chains and DEXes), and the unique value proposition (net profit calculations after bridge fees). This effectively distinguishes it from the generic 'arbitrage_scanner' sibling by emphasizing cross-chain analysis and fee-aware calculations.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No explicit guidance is provided on when to select this tool versus the 'arbitrage_scanner' sibling or other alternatives. While 'cross-chain' and 'bridge fees' imply usage for multi-chain opportunities, the description lacks explicit when-to-use/when-not-to-use criteria or references to alternative tools.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

crqc_proximity_benchmarkBInspect

CRQC Proximity Benchmark: Run standardized benchmark circuits on real IBM quantum hardware to measure current error rates and estimate how far we are from Cryptographically Relevant Quantum Computers (CRQC) that could break ECDSA.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full disclosure burden. It adds valuable context that the tool uses 'real IBM quantum hardware' (implying potential queue times and costs), but fails to specify execution time, cost implications, whether results are cached, or what the return format contains (error rates? qubit counts? time estimates?).

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single efficient sentence that front-loads the tool name and immediately explains its function. The colon structure ('CRQC Proximity Benchmark: Run...') is slightly redundant with the tool name but does not significantly impact clarity. Every word conveys specific meaning about the hardware platform, measurement type, and cryptographic goal.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given high complexity (quantum computing benchmarks, cryptographic relevance), absence of output schema, and lack of annotations, the description should explain return values, execution model (async vs sync), and cost structure. It provides none of these operational details, leaving significant gaps for an agent attempting to invoke the tool.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Input schema has 100% description coverage (chain, limit, timeperiod all documented). The description adds no explicit parameter guidance, so it earns the baseline score of 3. Notably, the description does not bridge the conceptual gap between running IBM quantum hardware benchmarks and the blockchain-query parameters present in the schema.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool runs standardized benchmark circuits on IBM quantum hardware to measure error rates and estimate proximity to CRQC (Cryptographically Relevant Quantum Computers) capable of breaking ECDSA. It distinguishes from siblings like 'pqc_readiness_assessment' (which assesses migration readiness) by focusing on raw hardware benchmarking and cryptographic breaking estimates.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides no guidance on when to use this tool versus sibling quantum tools like 'pqc_readiness_assessment', 'pqc_wallet_scan', or 'quantum_hardware_qrng'. It does not specify prerequisites (e.g., IBM Quantum account requirements) or when alternative analysis methods might be preferable.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

defi_yield_analysisBInspect

DeFi Yield Opportunity Finder: Scan DeFi protocols for highest risk-adjusted yield opportunities with gas cost analysis.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Since no annotations are provided, the description carries the full burden. It adds valuable behavioral context not in the schema: 'risk-adjusted' methodology and 'gas cost analysis'. However, it fails to disclose critical operational details: whether this is read-only (implied by 'scan' but not explicit), data sources, rate limits, or output format expectations.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Excellent structure—single sentence, front-loaded with functional identifier ('DeFi Yield Opportunity Finder'), followed by specific action verb and value propositions. Zero wasted words; every phrase (risk-adjusted, gas cost analysis) conveys distinct functionality.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given three simple optional parameters and no output schema, the description adequately covers the input side but remains incomplete regarding what the tool returns (e.g., yield percentages, risk metrics, protocol names, APY calculations). Without an output schema, the description should have sketched the return structure or analysis methodology.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

With 100% schema description coverage, the baseline is 3. The description mentions scanning 'DeFi protocols' which aligns with the 'chain' parameter, but adds no additional semantic context about how 'timeperiod' affects yield calculations (APY window vs. historical performance) or why one would adjust the 'limit' beyond the schema's default/max constraints.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool scans DeFi protocols to find yield opportunities, with specific differentiators (risk-adjusted yields, gas cost analysis). However, it lacks explicit differentiation from siblings like 'pendle_yield_trading' or 'flashloan_opportunities' that also deal with yield generation.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance provided on when to use this tool versus alternatives like 'arbitrage_scanner' or 'pendle_yield_trading'. No mention of prerequisites (e.g., specific chain selection requirements) or when this analysis is most valuable versus other investment research tools in the sibling list.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

einstein_meta_strategyCInspect

Einstein AI Meta-Strategy: AI-driven autonomous meta-strategy that decides what to trade across all asset classes (crypto swaps, Hyperliquid perps/stocks, Polymarket prediction markets, DeFi yield) on each execution cycle.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries full burden for behavioral disclosure but fails to specify whether the tool executes actual trades (destructive) or returns recommendations (read-only), what the return format is, or required authentication. For a financial trading tool, this is a critical safety gap.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Single sentence structure efficiently packs the conceptual scope. The 'Einstein AI Meta-Strategy:' prefix is slightly redundant with the tool name, but the sentence is front-loaded with the core value proposition and contains no extraneous filler.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the high complexity (autonomous AI trading across multiple asset classes), absence of output schema, and zero annotations, the description is insufficient. It omits execution semantics, return value structure, error handling, and whether human confirmation is required for the 'decisions' it makes.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, documenting 'chain', 'limit', and 'timeperiod'. The description lists asset classes covered by the strategy but does not explain how parameters filter or constrain the meta-strategy (e.g., whether 'chain' applies to all asset classes or only crypto swaps). Baseline score appropriate since schema carries the load.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states this is an 'AI-driven autonomous meta-strategy that decides what to trade' and enumerates specific asset classes (crypto swaps, Hyperliquid perps/stocks, Polymarket, DeFi yield). It distinguishes from siblings by positioning itself as a comprehensive 'meta' layer across all asset classes rather than a specialized scanner.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No explicit guidance is provided on when to invoke this meta-strategy versus using specific sibling tools like 'arbitrage_scanner' or 'defi_yield_analysis' directly. The phrase 'on each execution cycle' implies periodic use but doesn't clarify prerequisites or selection criteria.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

flashloan_opportunitiesBInspect

Flashloan Opportunity Scanner: Real-time scanner for MEV, liquidation, DEX arbitrage, and cross-chain flashloan opportunities with profit estimation.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full disclosure burden. It adds 'real-time' and 'profit estimation' as behavioral traits, but fails to clarify critical safety aspects: whether the tool executes transactions or is read-only, authentication requirements, rate limits, or data freshness guarantees. The term 'scanner' implies observation, but explicit safety context is missing for a high-risk financial domain.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, information-dense sentence with no filler content. It front-loads the tool identity and immediately lists specific capabilities, making efficient use of limited space.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the complexity of DeFi flashloans and the absence of both annotations and output schema, the description adequately covers the scanning scope but leaves gaps regarding return value structure, execution risks, and actionable next steps. It mentions profit estimation but does not characterize the opportunity data format.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The input schema has 100% description coverage with clear enum values and type definitions. The description does not add parameter-specific guidance (e.g., default behaviors, valid combinations) beyond what the schema already provides, warranting the baseline score of 3.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly identifies the tool as a scanner for flashloan opportunities, specifying four distinct opportunity types (MEV, liquidation, DEX arbitrage, cross-chain) and profit estimation. While it implicitly distinguishes from siblings like 'arbitrage_scanner' through specificity to flashloans, it does not explicitly contrast with similar tools in the ecosystem.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No explicit guidance is provided on when to use this tool versus similar siblings like 'arbitrage_scanner', 'cross_chain_arbitrage', or 'liquidation_monitor'. There are no prerequisites, exclusions, or workflow positioning hints, despite the crowded toolset with overlapping functionality.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

general_analysisAInspect

General AI Analysis: Send any prompt to Agent Einstein who autonomously selects tools, analyzes data, and provides intelligent responses. Use for complex questions requiring reasoning or multi-step analysis.

ParametersJSON Schema
NameRequiredDescriptionDefault
promptYesThe question or analysis request to send to Einstein AI
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description carries the full burden. It successfully discloses the crucial behavioral trait that this tool 'autonomously selects tools,' indicating it is an orchestrator/agent rather than a direct data source. It could improve by mentioning whether responses are streamed, cached, or have side effects, but the agentic nature is well-covered.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two efficient sentences with zero waste. The first sentence defines the mechanism and capability; the second provides usage guidance. Information is front-loaded and appropriately sized for a single-parameter tool.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the high complexity of an agent-orchestration tool among many specific function tools, the description adequately explains its role as the general interface for Einstein. It compensates for the lack of output schema by explaining the response type ('intelligent responses'). It appropriately omits redundant parameter detail given the 100% schema coverage.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The input schema has 100% description coverage for the single 'prompt' parameter ('The question or analysis request to send to Einstein AI'). The description text ('Send any prompt') aligns with but does not substantially extend the schema's semantics. With full schema coverage, the baseline score of 3 is appropriate.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool sends prompts to 'Agent Einstein' who autonomously selects tools and analyzes data. It distinguishes itself from the 40+ specific siblings (e.g., arbitrage_scanner, whale_tracking) by emphasizing this is a general-purpose, autonomous agent for complex queries rather than a single-function data fetch.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Provides clear positive guidance ('Use for complex questions requiring reasoning or multi-step analysis'), effectively positioning it as the fallback for queries that don't map to specific single-purpose siblings. Lacks explicit 'when not to use' exclusions or named alternatives, but the contrast with specific sibling names makes the boundary reasonably clear.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

holder_concentrationCInspect

Holder Concentration: Token holder distribution analysis with concentration risk metrics.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
tokenAddressNoToken contract address to analyze
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full burden of behavioral disclosure. It mentions 'concentration risk metrics' hinting at the output domain, but fails to disclose operational characteristics such as whether the analysis is real-time or cached, rate limits, error behaviors for invalid addresses, or what specific metrics (e.g., Gini coefficient, HHI) are returned.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness3/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single sentence, but wastes space by prefixing with 'Holder Concentration:' which restates the tool name. The remaining content 'Token holder distribution analysis with concentration risk metrics' is efficient, but the tautological opening reduces the value density.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a 4-parameter analytical tool with no output schema, the description provides the minimum viable context by identifying the analysis domain. However, it lacks detail on the return format or specific concentration metrics calculated, leaving significant gaps given the absence of structured output documentation.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Input schema has 100% description coverage, establishing a baseline of 3. The description mentions 'concentration risk metrics' which contextually implies the 'tokenAddress' is the primary subject, but adds no specific guidance on parameter usage (e.g., syntax for addresses, how 'limit' affects the depth of holder analysis, or default behaviors when optional parameters are omitted).

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool performs 'Token holder distribution analysis with concentration risk metrics,' specifying the resource (token holders) and output type (risk metrics). It implicitly distinguishes from sibling tools like 'whale_tracking' (individual movements) and 'wallet_portfolio' (single wallet views) through its focus on aggregate distribution and concentration risk, though it doesn't explicitly name alternatives.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides no guidance on when to use this tool versus alternatives. It doesn't indicate prerequisites (e.g., needing a specific token address format), when to prefer this over 'whale_tracking' for large holder analysis, or exclusions for certain token types.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

hyperliquid_smart_moneyBInspect

Hyperliquid Smart Money Intelligence: Access top trader leaderboard, whale positions, and smart money consensus signals from Hyperliquid. Shows what the best traders are collectively bullish or bearish on.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full burden of behavioral disclosure. It describes the semantic content of returned data (bullish/bearish signals, leaderboards) but omits operational characteristics such as real-time vs cached data, rate limits, authentication requirements, or safety properties (read-only vs destructive).

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description consists of two efficient sentences with no redundancy. The first sentence front-loads the core capability and data sources, while the second clarifies the value proposition (collective bullish/bearish sentiment), making every word serve a distinct purpose.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given three optional parameters with full schema documentation and no output schema, the description adequately covers the business purpose. However, without annotations or an output schema, the description should ideally characterize the return structure (e.g., 'returns a list of traders with PnL rankings') to be complete.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, establishing a baseline of 3. The description provides conceptual context about what data is retrieved (implying the 'limit' applies to traders or signals) but does not explicitly map parameters to functionality, provide usage examples, or explain parameter interactions (e.g., chain filtering on Hyperliquid).

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly identifies the tool's function as accessing Hyperliquid-specific smart money data including leaderboards, whale positions, and consensus signals. It distinguishes the scope to Hyperliquid (differentiating from the generic 'smart_money' sibling) by explicitly mentioning the platform twice, though it doesn't explicitly contrast usage against the sibling tool.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No explicit guidance is provided for when to select this tool versus the generic 'smart_money' sibling or other trading intelligence tools like 'copy_trading_intel'. The description implies Hyperliquid-specific use cases through naming but lacks explicit when-to-use or when-not-to-use instructions.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

liquidation_monitorAInspect

DeFi Liquidation Monitor: Monitor Aave/Compound positions approaching liquidation with health factors and bonus estimates.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries full burden of behavioral disclosure. It successfully indicates the tool returns 'health factors and bonus estimates' (domain-specific outputs), and implies read-only monitoring behavior. However, it lacks explicit safety confirmation (read-only vs transactional), error behavior, or real-time vs cached data disclosure that would be critical for financial DeFi operations.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Single sentence format with category prefix 'DeFi Liquidation Monitor:' provides immediate categorization. Every clause delivers distinct value: protocol scope (Aave/Compound), function (monitoring liquidation proximity), and return data types (health factors, bonus estimates). Zero redundancy or filler.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given 100% schema coverage and no output schema, the description adequately compensates by specifying what data is returned (health factors, bonus estimates). However, it misses opportunity to note that all parameters are optional or describe default monitoring behavior when no parameters are provided.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, establishing baseline 3. The description mentions Aave/Compound contexts but does not add parameter-specific semantics beyond the schema (e.g., default behaviors, valid combinations, or that all parameters are optional). No clarification on whether 'chain' filters protocols or just networks.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Description uses specific verb 'Monitor' with explicit resources 'Aave/Compound positions' and clarifies scope with 'approaching liquidation'. The mention of 'health factors and bonus estimates' distinguishes it from sibling tools like defi_yield_analysis or flashloan_opportunities, clearly positioning it as a liquidation risk assessment tool.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Description provides no explicit guidance on when to use this tool versus alternatives (e.g., when to use flashloan_opportunities vs this monitor). No prerequisites, filtering guidance, or 'when-not-to-use' conditions are specified despite the tool having specific DeFi protocol limitations.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

market_moversCInspect

Market Movers: Detect significant price action and momentum signals across DEXes.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description carries the full burden. While 'Detect' implies a read-only operation, the description doesn't confirm safety, disclose rate limits, define what constitutes 'significant' price action, or indicate data freshness/latency.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Single sentence efficiently delivers the core purpose without redundancy. The 'Market Movers:' prefix substitutes for the null title effectively. No filler content.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a complex financial analysis tool with no output schema and no annotations, the description is insufficient. It fails to explain the return format, what specific momentum indicators are used, or how results are ranked, leaving the agent unprepared to interpret results.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Input schema has 100% description coverage with clear enums for chain and timeperiod. The description adds minimal semantic value beyond the schema, though the mention of 'across DEXes' implicitly contextualizes the chain parameter. Baseline 3 is appropriate given schema completeness.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool detects 'significant price action and momentum signals across DEXes' with specific verbs and domain. However, it lacks explicit differentiation from siblings like 'top_gainers' or 'technical_analysis' which operate in similar domains.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance provided on when to use this tool versus the 40+ sibling analysis tools (e.g., 'top_gainers', 'smart_money', 'whale_tracking'). No prerequisites, filters, or exclusion criteria are mentioned.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

mev_detectionCInspect

MEV Detection: Sandwich attacks, front-running, and toxic flow analysis.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full burden of behavioral disclosure. While it lists attack types analyzed, it fails to disclose whether the operation is read-only, what data format is returned, rate limits, or whether results are real-time vs historical. Insufficient for a complex blockchain analytics tool.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness3/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is extremely terse (7 words) and front-loaded, but undersized for the tool's complexity. While no words are wasted, critical information about return values and sibling differentiation is missing, making it inadequately concise rather than efficiently minimal.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the complexity of MEV analysis and absence of an output schema, the description should explain what data structure is returned (transactions, statistics, scores) and analysis methodology. With 3 optional parameters and no return documentation, the description is incomplete for safe invocation.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, establishing a baseline of 3. The description mentions specific MEV attack types which provides domain context, but adds no specific guidance on parameter interactions (e.g., default behavior when all optional parameters are omitted) or valid value formats beyond the schema.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly identifies the tool's function using specific verbs ('Detection') and resources ('Sandwich attacks, front-running, toxic flow'), which helps distinguish it as an analysis tool. However, it does not explicitly differentiate from the sibling 'mev_protection' tool, which could cause selection confusion.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance provided on when to use this tool versus alternatives like 'mev_protection' or 'arbitrage_scanner'. No mention of prerequisites, data freshness, or whether this requires specific blockchain node access.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

mev_protectionCInspect

MEV Protection Transaction Routing: Analyze transactions for MEV vulnerability and recommend private mempool routing strategies.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries full burden but remains ambiguous on critical behavioral traits: it is unclear whether this tool actually performs transaction routing or merely recommends strategies, what the analysis criteria are, rate limits, or required permissions for mempool access.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Single sentence with subtitle structure. The prefix 'MEV Protection Transaction Routing:' slightly wastes space by restating the concept implicit in the tool name, but the remainder efficiently communicates core functionality without redundancy.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

With 3 optional parameters and no output schema, the description should clarify what default behavior occurs when called without arguments. It omits explanation of return values, analysis methodology, or how the 'recommendations' are formatted, leaving significant gaps for an agent attempting to use the tool effectively.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100% (chain, limit, timeperiod all documented). The description adds no additional parameter context, but meets the baseline since the schema fully documents acceptable values (enums) and defaults.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description uses specific verbs ('analyze', 'recommend') and identifies the domain (MEV vulnerability, private mempool routing). However, it fails to distinguish from sibling tool 'mev_detection', leaving ambiguity about whether this tool detects threats, provides protection analysis, or executes routing.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance provided on when to use this versus 'mev_detection' or other analysis tools. No mention of prerequisites (e.g., transaction data requirements) or when private mempool routing is actually necessary versus standard broadcasting.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

pendle_yield_tradingBInspect

Pendle Yield Tokenization Analysis: Analyze Pendle yield markets, compare fixed vs variable yields, find best PT/YT opportunities across chains, and view yield term structures.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, placing full burden on the description. It discloses analytical behaviors ('compare fixed vs variable yields', 'view yield term structures') but omits operational details such as whether the tool is read-only, rate limits, data freshness, or authentication requirements.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single compound sentence that efficiently lists multiple capabilities. It is appropriately sized and front-loaded with the domain identifier ('Pendle Yield Tokenization Analysis'), though the colon-separated label format slightly reduces readability.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

The description adequately covers the tool's domain-specific purpose (Pendle yield tokenization) and analytical capabilities. However, it lacks description of return values (no output schema exists) and operational constraints, leaving gaps for a tool with financial analysis capabilities.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, providing clear descriptions for 'chain', 'limit', and 'timeperiod'. The description mentions 'across chains' which aligns with the chain parameter, but does not add semantic meaning or usage guidance beyond what the schema already provides. Baseline 3 is appropriate for high schema coverage.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description uses specific action verbs ('Analyze', 'compare', 'find', 'view') and identifies the specific resource ('Pendle yield markets', 'PT/YT opportunities'). It distinguishes from siblings like 'defi_yield_analysis' by explicitly referencing Pendle-specific concepts (Principal/Yield Tokens, fixed vs variable yields).

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides no explicit guidance on when to use this tool versus the sibling 'defi_yield_analysis' or other yield-related tools. While the Pendle specificity implies usage for Pendle markets, there is no explicit 'when to use' or 'when not to use' guidance.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

polymarket_arbitrageAInspect

Polymarket Cross-Platform Arbitrage: Find price discrepancies between Polymarket and Kalshi prediction markets for arbitrage opportunities with confidence scoring.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full disclosure burden. It adds valuable behavioral context by mentioning 'confidence scoring' as a feature, but fails to clarify critical operational traits: whether the tool executes trades or merely reports opportunities, authentication requirements, or the confidence score range/format.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, efficient sentence with zero redundancy. It front-loads the domain ('Polymarket Cross-Platform Arbitrage') and immediately follows with the mechanism and distinctive feature. Every clause serves a specific purpose in defining the tool's unique value proposition.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the 100% schema coverage and absence of an output schema, the description adequately covers the tool's purpose but leaves gaps regarding return value structure (e.g., whether it returns opportunities or execution results). For a financial analysis tool with three well-documented parameters, it is minimally viable but could benefit from describing the output format.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The input schema has 100% description coverage with clear enums and default values documented. The description does not add semantic meaning beyond the schema (e.g., it doesn't explain why one would choose a specific chain or timeperiod), but with complete schema documentation, it meets the baseline expectation.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the specific action (find price discrepancies), the exact platforms compared (Polymarket and Kalshi), and the output characteristic (confidence scoring). It effectively distinguishes from siblings like polymarket_spread_analysis (internal spreads) and cross_chain_arbitrage (crypto chains) by explicitly naming the Kalshi cross-platform comparison.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

While the description implies usage scenarios through the specificity of 'Cross-Platform' and naming Kalshi, it lacks explicit guidance on when to prefer this tool over arbitrage_scanner or polymarket_conviction. No 'when-not-to-use' or prerequisite conditions are mentioned, leaving agents to infer applicability.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

polymarket_convictionAInspect

Polymarket High Conviction Scanner: Identify high-conviction betting opportunities by analyzing model-price divergence, cross-platform comparisons, and order flow signals.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full burden of behavioral disclosure. While it mentions the analytical approach (model-price divergence, order flow), it omits operational traits: whether results are real-time or cached, rate limiting, authentication requirements, or whether the operation is read-only. This leaves significant behavioral gaps.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is optimally concise with a clear header ('Polymarket High Conviction Scanner:') followed immediately by the action verb and methodology. Zero redundant words; every clause serves to differentiate the tool's specific value proposition.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool has no output schema and no annotations, the description adequately covers the functional purpose but lacks completeness regarding return values or result structure. For a 3-parameter analysis tool, it meets minimum viability but should disclose what constitutes a 'high-conviction opportunity' in the output.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, with all three parameters (chain, limit, timeperiod) fully documented in the schema. The description does not add semantic value beyond the schema (e.g., no examples of valid combinations or clarification on default behaviors), warranting the baseline score.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool identifies 'high-conviction betting opportunities' using specific methodologies (model-price divergence, cross-platform comparisons, order flow signals). It effectively distinguishes itself from sibling tools like polymarket_arbitrage and polymarket_insider_scan through its unique focus on conviction signals versus pure arbitrage or insider detection.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage context (when seeking high-conviction signals with model-price divergence), but provides no explicit when-to-use guidance or comparisons to alternatives. It does not clarify why one would choose this over polymarket_arbitrage or polymarket_insider_scan despite having distinct functionality.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

polymarket_insider_scanAInspect

Polymarket Insider Detection: Scan Polymarket for insider trading signals using 18 detection heuristics including wallet clustering, Sybil detection, and on-chain whale transfer analysis.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full disclosure burden. It successfully describes the analytical behavior (18 heuristics, specific detection methods) but fails to disclose operational traits like read-only status, rate limits, authentication requirements, or output format. It confirms what analysis is performed but not safety profile or side effects.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Single dense sentence with zero wasted words. Front-loaded with the core function ('Polymarket Insider Detection: Scan...'), immediately communicating value. Every phrase ('18 detection heuristics', 'wallet clustering', 'Sybil detection') adds specific methodological context without redundancy.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Adequate for purpose identification but incomplete given the tool's complexity and lack of output schema. For an analysis tool running 18 heuristics, the description should ideally characterize the output format (alerts? scores? clustered wallets?) or signal types returned. It mentions 'signals' but leaves significant gaps regarding the return structure.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, establishing a baseline of 3. The description adds no parameter-specific guidance (e.g., no mention of tokenAddress, timeperiod granularity implications, or chain selection criteria), relying entirely on the schema for parameter documentation.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Description clearly states the tool scans Polymarket for insider trading signals, specifying the exact resource (Polymarket) and action (scan/detection). The mention of '18 detection heuristics' and specific methodologies (wallet clustering, Sybil detection, whale transfer analysis) strongly distinguishes it from sibling tools like polymarket_arbitrage or polymarket_markets.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

While it lacks explicit 'when-not-to-use' language, the description provides clear contextual boundaries through the specific detection heuristics listed. An agent can easily infer this tool is for insider/suspicious activity detection rather than general market analysis or arbitrage, effectively differentiating it from the 9+ sibling Polymarket tools without explicit enumeration.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

polymarket_marketsCInspect

Polymarket Market Discovery: Search and discover active prediction markets on Polymarket with AI-powered semantic search, order book data, and analytics.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full burden of behavioral disclosure. While it mentions return data types (order book, analytics), it misleadingly implies semantic search functionality that the parameters cannot support. It lacks disclosure on authentication requirements, rate limits, read-only status, or pagination behavior beyond the limit parameter.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, efficient sentence with no obvious waste. The prefix 'Polymarket Market Discovery:' is slightly redundant with the tool name but serves as a subject header. The content is front-loaded with the key resource and action.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a tool with 3 optional parameters and no output schema, the description is minimally adequate. It compensates for the missing output schema by mentioning return data types (order book, analytics). However, the unsupported 'semantic search' claim represents a significant gap in accuracy, and the lack of sibling differentiation leaves the tool's specific role unclear.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, establishing a baseline of 3. The description does not add specific parameter semantics beyond the schema (e.g., valid chain values, time period formats), and the 'semantic search' claim creates confusion about how the chain/limit/timeperiod parameters function (as filters vs. search inputs).

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose3/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description identifies the resource (Polymarket prediction markets) and general action (search/discover), mentioning specific return data like order book and analytics. However, it claims 'AI-powered semantic search' capabilities that are unsupported by the input schema (which lacks any query/text parameter), and fails to distinguish this tool from four siblings (polymarket_arbitrage, polymarket_conviction, polymarket_insider_scan, polymarket_spread_analysis).

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No explicit when-to-use guidance is provided, nor are alternatives named. Given the four specialized Polymarket siblings, the description misses a critical opportunity to clarify that this is the general discovery tool while siblings handle specific analyses (arbitrage, conviction, etc.).

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

polymarket_spread_analysisBInspect

Polymarket Spread & Market Making: Analyze bid-ask spreads, completeness arbitrage (YES+NO < $1), and market making opportunities with real CLOB order book data.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description adds valuable behavioral context: it specifies the data source ('real CLOB order book data') and the specific arbitrage logic ('completeness arbitrage (YES+NO < $1)'). However, it lacks disclosure on rate limits, authentication requirements, or what constitutes a 'market making opportunity' in the output.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Excellent information density in a single sentence. The colon-separated structure front-loads the domain ('Polymarket Spread & Market Making') followed by specific analytical capabilities. Every phrase earns its place without redundancy.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

No output schema exists, yet the description fails to indicate return value structure (e.g., spread metrics, arbitrage opportunities, or order book depth). With 3 optional parameters and no annotations, the description should also explain default behavior when no parameters are provided, which it does not.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, establishing a baseline of 3. The description mentions 'real CLOB order book data' which loosely implies blockchain context, but does not explicitly reference chain, limit, or timeperiod parameters, nor does it explain default behaviors for these optional parameters.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool analyzes bid-ask spreads, completeness arbitrage (YES+NO < $1), and market making opportunities using real CLOB order book data. It effectively distinguishes from generic polymarket_arbitrage by specifying 'completeness arbitrage' and 'market making,' though it doesn't explicitly contrast use cases with siblings.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No explicit guidance on when to use this versus polymarket_arbitrage or other Polymarket siblings. While 'market making opportunities' implies usage for liquidity providers, there are no stated prerequisites, exclusions, or decision criteria for tool selection.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

pqc_migration_plannerAInspect

PQC Migration Planner: Generate a prioritized post-quantum migration plan for wallet addresses. Ranks by exposure (risk × balance), provides step-by-step migration instructions with destination recommendations and gas cost estimates.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full burden. It adds valuable context about the ranking algorithm (risk × balance) and output components (step-by-step instructions, gas estimates), but fails to clarify whether this tool executes transactions or is read-only analysis, and does not explain how wallet addresses are determined given the lack of a wallet address parameter in the schema.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is optimally concise with two high-density sentences. The first front-loads the core purpose (Generate a prioritized post-quantum migration plan), while the second efficiently enumerates the specific deliverables (ranking methodology, instructions, recommendations, estimates) without redundancy.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given no output schema and no annotations, the description adequately explains what the generated plan contains conceptually, but has notable gaps: it does not clarify how wallet addresses are specified (absent from schema parameters), lacks safety/destructive behavior disclosure, and does not describe the return format structure despite the absence of an output schema.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, establishing a baseline of 3. The description does not add additional semantic context for chain, limit, or timeperiod beyond what the schema already provides, nor does it explain the relationship between these parameters and the 'wallet addresses' mentioned in the description.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the specific action (generate a prioritized post-quantum migration plan), the target resource (wallet addresses), and distinguishes from siblings like pqc_readiness_assessment (assessment) and pqc_wallet_scan (scanning) by emphasizing actionable planning outputs (step-by-step instructions, destination recommendations, gas estimates).

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

While the description implies usage through specificity (generating actionable plans vs. mere assessment), it lacks explicit guidance on when to select this over pqc_readiness_assessment or pqc_wallet_scan. No 'when to use' or explicit alternative recommendations are provided.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

pqc_readiness_assessmentCInspect

PQC Crypto Readiness Assessment: Comprehensive ecosystem-wide quantum readiness report covering chain PQC status, NIST algorithm comparison (Dilithium, FALCON, SPHINCS+), quantum threat timeline, and optional portfolio exposure analysis.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full burden of behavioral disclosure. It fails to indicate whether this is a read-only operation (highly likely given 'assessment'), whether it requires authentication, potential rate limits for quantum analysis, or the expected response format/structure. It only describes output topics, not behavioral traits or side effects.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is efficiently structured as a title header followed by a comprehensive feature list. Every clause adds specific content information (algorithm names, analysis types). It avoids fluff, though the colon-separated title prefix slightly echoes the tool name without adding new information.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the absence of an output schema, the description partially compensates by listing report contents (algorithms, timeline, portfolio exposure). However, it omits the return structure (JSON vs markdown), whether the analysis is synchronous or async, and how the 'optional' portfolio analysis is triggered (no corresponding boolean parameter exists in the schema).

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, providing clear definitions for chain, limit, and timeperiod. The description adds marginal context by linking 'chain' to 'chain PQC status' but does not explain parameter interactions (e.g., whether timeperiod affects the portfolio analysis or just the threat timeline). With the schema doing the heavy lifting, this meets the baseline.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states it generates a 'quantum readiness report' and lists specific content areas (NIST algorithms like Dilithium/FALCON/SPHINCS+, threat timelines, chain status). The technical specificity distinguishes it from sibling tools like pqc_migration_planner (which implies action) and pqc_wallet_scan (which implies targeted scanning), though it could more explicitly state this is an analytical/reporting tool rather than an operational one.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

There is no guidance on when to select this tool versus the related pqc_migration_planner, pqc_wallet_scan, or quantum_portfolio_monitor. The phrase 'optional portfolio exposure analysis' hints at a use case but doesn't clarify selection criteria or prerequisites (e.g., whether a specific chain must be selected for portfolio analysis to function).

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

pqc_wallet_scanBInspect

Quantum Address Resistance Scanner: Deep analysis of any Bitcoin or Ethereum wallet address for quantum resistance. Detects address type (P2PKH, P2SH, P2WPKH, P2TR, EOA, Contract), assesses vulnerability to Shor's and Grover's algorithms, and provides expert PQC migration advice.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description carries the full burden. It discloses the analysis scope (detects address types, assesses vulnerabilities) and mentions specific quantum algorithms, but fails to describe output format, whether results are cached, rate limits, or if the tool makes external blockchain queries.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two efficient sentences with zero filler. The first acts as a descriptive title, the second lists specific technical capabilities. All content earns its place, though the leading label 'Quantum Address Resistance Scanner:' partially restates the tool name.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a complex quantum cryptography tool with 3+ parameters and no output schema, the description adequately explains the analysis purpose but lacks detail on return values or execution requirements. Given the technical complexity (Shor's algorithm assessment), more behavioral context would be appropriate.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100% for the three defined parameters (chain, limit, timeperiod), establishing a baseline of 3. The description implies an address input ('wallet address') which aligns with the required 'tokenAddress' field, but does not explain how 'limit' or 'timeperiod' apply to quantum resistance analysis (e.g., transaction history analysis vs. static address assessment).

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool performs 'deep analysis' for 'quantum resistance' on wallet addresses, specifying exact address types (P2PKH, P2SH, P2WPKH, P2TR, EOA, Contract) and algorithms (Shor's, Grover's). However, it does not distinguish from siblings like 'pqc_migration_planner' or 'pqc_readiness_assessment' which also deal with quantum cryptography.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance provided on when to use this tool versus alternatives like 'pqc_migration_planner' (which offers migration advice, similar to this tool's 'PQC migration advice') or 'pqc_readiness_assessment'. No prerequisites or exclusion criteria are mentioned.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

pump_fun_launchesAInspect

Pump.fun Launches: Fresh token launches on Pump.fun with bonding curve progress (Solana only).

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full burden. It adds valuable context about 'bonding curve progress' (specific to Pump.fun mechanics) and the Solana chain constraint, but omits rate limits, data freshness definitions, or read-only/destructive nature.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Single sentence with minimal waste. The leading 'Pump.fun Launches:' is slightly redundant with the tool name but serves as effective framing. Key constraints (Solana only, bonding curves) are front-loaded.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Adequate for a data retrieval tool with full schema coverage and no output schema. Captures the domain-specific context (Pump.fun, bonding curves) necessary for an agent to understand what data is returned, though could specify data freshness/recency.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Despite 100% schema description coverage, the description adds critical semantic value by stating 'Solana only'—directly contradicting the schema's permissive enum (base, ethereum, etc.) to clarify the actual valid input for the chain parameter.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

States it retrieves fresh token launches from Pump.fun with bonding curve progress, and specifies 'Solana only' which distinguishes it from sibling tools like zora_launches or general EVM tools. Lacks an explicit verb but the resource and scope are clear.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The 'Solana only' constraint implies when to use it (Solana/Pump.fun specific queries), but provides no explicit alternatives or when-not-to-use guidance versus siblings like token_sniping or general market_movers.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

quantum_encryption_keyDInspect

Quantum Encryption Key Generator: Generate AES-256 encryption keys from real ANU quantum entropy. Returns 256-bit key in hex and base64 formats.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Without annotations, the description carries full disclosure burden. It mentions output formats (hex and base64) and entropy source (ANU), which is helpful if accurate. However, given the parameter mismatch, this behavioral information is likely incorrect or describes a different endpoint.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is efficiently structured with a descriptive prefix and single explanatory sentence. No verbose filler, though the first clause acts as a title rather than content.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness1/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Critical failure: the description does not reconcile the stated purpose (key generation) with the actual parameters (blockchain querying). A complete description would explain the relationship between timeperiod/limit and key generation, or acknowledge this as a blockchain analytics tool.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters2/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Despite 100% schema description coverage (baseline 3), the description fails to mention the actual parameters (chain, limit, timeperiod) or explain why a key generator requires blockchain network and time range filters. The domain mismatch actively confuses parameter semantics.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose2/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states it generates AES-256 encryption keys from quantum entropy, but this appears to describe a different tool entirely. The input parameters (chain, limit, timeperiod) clearly indicate a blockchain data querying function, not key generation, making the description actively misleading.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines1/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance provided on when to use this tool versus quantum-related siblings like quantum_signing_key, quantum_hardware_qrng, or quantum_wallet_generation. No mention of prerequisites or when-not-to-use constraints.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

quantum_hardware_qrngDInspect

Hardware-Verified Quantum Random Number Generator: Generate cryptographically secure random numbers on real IBM quantum hardware (127-qubit superconducting processor). Each result includes IBM job ID as proof of quantum origin.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Without annotations, the description carries the full burden of behavioral disclosure. It mentions IBM job IDs as proof of origin, which adds value, but fails entirely to explain the blockchain-related parameters (chain, timeperiod) or reconcile them with the quantum RNG purpose.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is efficiently structured with two focused sentences that front-load the key value proposition (quantum hardware verification). However, the content is inappropriate for the actual interface.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness1/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the severe mismatch between the described quantum RNG functionality and the blockchain query parameters, the description is fundamentally incomplete. It provides no context for the chain, limit, or timeperiod parameters, rendering the tool interface unexplained.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters1/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

While the schema has 100% coverage (baseline 3), the description implies quantum-specific parameters (qubits, shots, backend) but the actual parameters are for blockchain queries (chain, timeperiod, limit). This contradiction creates confusion rather than adding meaning, warranting a score of 1.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose2/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states it generates quantum random numbers on IBM hardware, which distinguishes it from sibling tools. However, this purpose directly contradicts the input schema, which accepts blockchain network parameters (ethereum, solana, etc.) and time periods, suggesting the tool is actually for blockchain analytics rather than quantum RNG.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines1/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance is provided on when to use this tool versus alternatives. Given the parameter schema suggests blockchain querying while the description claims quantum random number generation, agents have no way to determine correct usage or prerequisites.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

quantum_portfolio_monitorCInspect

Quantum Portfolio Risk Monitor: Scan multiple wallet addresses across chains for quantum vulnerability. Aggregates risk scores, identifies value at risk, and provides portfolio-level quantum threat assessment with risk distribution.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full burden of behavioral disclosure. It successfully describes expected outputs (risk scores, value at risk, threat assessment, risk distribution) but fails to clarify operational aspects like whether this is read-only, rate limits, or how the 'multiple wallet addresses' are determined given the lack of address parameters.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is efficiently structured in two sentences with minimal waste, though the prefix 'Quantum Portfolio Risk Monitor:' restates the tool name and could be removed. The content is appropriately front-loaded with the core action before listing specific outputs.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given no output schema exists, the description adequately compensates by detailing return value types (risk scores, distributions). However, the unexplained discrepancy between the wallet-scanning description and the address-less input schema leaves critical context gaps regarding how the tool identifies which portfolios to monitor.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters2/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

While schema description coverage is 100% (baseline 3), the description introduces confusion by referencing 'multiple wallet addresses' and 'across chains' when the schema only accepts a single chain enum and provides no wallet address input field. This mismatch between described functionality and actual parameters reduces the score.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose3/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description provides specific verbs (Scan, Aggregates, identifies) and mentions the resource (wallet addresses, portfolio-level assessment). However, it fails to differentiate from sibling tools like 'pqc_wallet_scan' or 'pqc_readiness_assessment', and critically, it claims to scan wallet addresses while the input schema provides no parameter to specify them.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance is provided on when to use this tool versus alternatives like 'pqc_wallet_scan' (individual vs. portfolio) or 'pqc_migration_planner'. There are no prerequisites mentioned despite the complex quantum cryptography domain, and no warnings about the 0 required parameters potentially indicating global chain-wide scanning behavior.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

quantum_signing_keyCInspect

Quantum Signing Key Generator: Generate Ed25519 or secp256k1 signing keypairs from real ANU quantum entropy. Returns private key and public key pair.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Without annotations, the description carries full burden. It discloses the entropy source (ANU quantum) and return values (private/public key pair), but omits critical behavioral details: key format encoding, whether keys are ephemeral or stored, and security warnings for handling private key material.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two concise sentences with efficient front-loading. Structure is appropriate, though the content creates potential confusion given the schema mismatch.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given no output schema, the description minimally covers return values. However, it fails to explain the parameter mismatch (why a key generator takes 'timeperiod' and 'limit'), omits security warnings critical for key generation tools, and doesn't clarify key formatting or encoding.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters1/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Severe mismatch between description and schema. While the description implies curve selection (Ed25519/secp256k1), the schema provides analytics parameters (chain, limit, timeperiod). The description adds no meaning to these specific parameters, and 'limit'/'timeperiod' are nonsensical for key generation as described.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose3/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description states a clear purpose (generate Ed25519/secp256k1 keypairs from quantum entropy) with specific cryptographic resources. However, it fails to align with the input schema which contains analytics parameters (limit, timeperiod), creating confusion about actual functionality.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance provided on when to use this tool versus siblings like quantum_wallet_generation or quantum_encryption_key. No mention of prerequisites for key generation or security considerations for handling private keys.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

quantum_wallet_generationAInspect

Quantum Wallet Generation: Generate BIP-39 mnemonic seed phrases from real ANU quantum entropy (vacuum fluctuation measurements). Returns 12 or 24-word mnemonics with derived addresses for EVM, Solana, and Bitcoin.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full burden. It successfully discloses the entropy source (ANU vacuum fluctuation measurements) and output format (mnemonics + derived addresses for EVM/Solana/Bitcoin). It does not address whether wallets are persisted server-side, idempotency, or rate limits.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is efficiently structured as a single informative sentence. It loses one point for the redundant prefix 'Quantum Wallet Generation:' which restates the tool name, though the core content is dense and valuable.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Critical gaps exist: no output schema is present, no annotations are provided, and the description does not explain why a wallet generation tool accepts 'timeperiod' and 'limit' parameters (which suggest analytics/monitoring functionality). For a cryptographic tool handling seed phrases, the lack of security handling disclosure and parameter rationale is a significant omission.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, establishing a baseline of 3. The description does not add parameter-specific context, but this is acceptable per rubric guidelines. Notably, the description fails to reconcile the stated wallet generation purpose with the query-like parameters ('limit', 'timeperiod') present in the schema.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool generates BIP-39 mnemonic seed phrases using ANU quantum entropy, and specifies the output (12/24-word mnemonics with derived addresses). It effectively distinguishes from siblings like 'quantum_encryption_key' and 'quantum_signing_key' by specifying wallet-specific BIP-39 mnemonics rather than generic cryptographic keys.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description explains what the tool does but provides no guidance on when to use it versus alternatives (e.g., when to use this versus 'pqc_migration_planner' or regular wallet generation). It lacks prerequisites, security warnings, or exclusion criteria for when NOT to use it.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

runes_analyticsBInspect

Runes Protocol Analytics: Runes protocol overview, mint progress, holder distribution, and individual rune details. Includes mintable status and etching data.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full disclosure burden. It compensates somewhat by detailing what data is returned (overview, holder distribution, mint status), but omits operational details like rate limits, data freshness, or caching behavior that would help an agent understand invocation constraints.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two sentences with minimal bloat, though the opening 'Runes Protocol Analytics: Runes protocol...' is slightly redundant. The information is front-loaded with the protocol name and specific data categories, efficiently communicating value despite the repetition.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the lack of output schema, the description appropriately lists return value categories (mint progress, holder distribution, etc.). However, it misses the opportunity to clarify the Bitcoin-specific nature of Runes relative to the multi-chain input schema, leaving a significant gap in contextual understanding for an agent selecting parameters.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, so the baseline is 3. The description adds no explicit parameter guidance (e.g., it doesn't clarify that 'chain' likely should be Bitcoin for Runes data, despite the broad enum), but the schema itself adequately documents the parameters without requiring additional description support.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly identifies the tool's focus on 'Runes Protocol Analytics' and lists specific data points returned (mint progress, holder distribution, etching data, mintable status). It distinguishes from siblings like 'brc20_analytics' and 'bitcoin_onchain_analytics' by explicitly naming the Runes protocol, though it could strengthen this by clarifying it's specifically for Bitcoin Runes.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No explicit guidance on when to use this versus 'brc20_analytics' or other Bitcoin analytics tools. Critically, it fails to address that Runes is a Bitcoin-specific protocol while the input schema allows non-Bitcoin chains (Ethereum, Solana, etc.), leaving users unclear about valid parameter combinations.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

security_scanBInspect

Token Security Scanner: Rug pull detection and security analysis with risk scoring.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
tokenAddressYesToken contract address to analyze
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full burden of behavioral disclosure. It successfully indicates the analysis type (rug pull detection, risk scoring) but omits operational details like whether this is read-only, rate limits, data sources, or what the risk score format/scale looks like.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single efficient sentence with no redundancy, though the prefix 'Token Security Scanner:' partially restates the tool name. It is front-loaded with the key capability (rug pull detection). However, its extreme brevity leaves gaps given the lack of annotations and output schema.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the absence of annotations and output schema, the description inadequately explains what the tool returns (e.g., risk score format, specific indicators checked). While it mentions 'risk scoring', a security tool with 4 parameters and no structured output documentation requires more descriptive context to be fully usable.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The input schema has 100% description coverage for all 4 parameters, establishing a baseline of 3. The description itself does not mention any parameters or provide additional context about valid address formats or default behaviors beyond what the schema already documents.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly specifies the resource (tokens) and actions (rug pull detection, security analysis, risk scoring), providing concrete behavioral details beyond the tool name. However, it does not distinguish this from siblings like 'smart_contract_audit' or 'wallet_trust_score'.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides no guidance on when to use this tool versus alternatives like 'smart_contract_audit', nor does it specify prerequisites (e.g., valid token address format) or exclusion criteria. It assumes the user knows when a security scan is appropriate.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

smart_contract_auditCInspect

Smart Contract Security Audit: Comprehensive smart contract vulnerability scanning across 30+ patterns with AI-powered reports and remediation suggestions.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full burden but only discloses that it generates 'AI-powered reports and remediation suggestions'. It fails to clarify whether the tool performs active mutation (deploying test transactions), whether it is asynchronous, or how to specify the target contract given the schema lacks a contract address field.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single efficient sentence that front-loads the function ('Smart Contract Security Audit:') before detailing capabilities. It avoids redundancy with the tool name while conveying key deliverables without excess verbiage.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the schema lacks a contract identifier parameter yet the description implies scanning functionality, the description is incomplete regarding what exactly gets audited (e.g., recent deployments, specific addresses). Without an output schema, it should better clarify the return structure beyond 'reports'.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, documenting chain, limit, and timeperiod. The description adds no specific parameter semantics beyond the schema, though it implies the scope of analysis ('30+ patterns') which does not map to specific inputs. Baseline score applies.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool performs 'Smart Contract Security Audit' with 'vulnerability scanning across 30+ patterns' and specifies outputs like 'AI-powered reports'. This distinguishes it from the generic `security_scan` sibling by focusing on smart contracts specifically. However, it does not explicitly differentiate usage contexts from similar security tools.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides no guidance on when to select this tool versus alternatives like `security_scan` or other analysis tools. It lacks prerequisites (e.g., whether a contract address is required) and does not indicate if it is for pre-deployment or post-deployment analysis.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

smart_moneyBInspect

Smart Money Analysis: Analyze top trader wallets with flow analysis and performance metrics.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
walletAddressNoWallet address to analyze
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full burden of behavioral disclosure. It specifies the analysis types ('flow analysis', 'performance metrics') but omits critical operational details: whether the data is real-time or cached, what defines 'smart money' in this context (profitability thresholds, volume, etc.), and whether the operation is read-only (implied but not stated).

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single efficient sentence fragment, but the prefix 'Smart Money Analysis:' restates the tool name and wastes valuable context space. The core content ('Analyze top trader wallets with flow analysis and performance metrics') is appropriately dense and front-loaded.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given four optional parameters with complete schema coverage and no output schema, the description meets minimum viability but is minimal for the domain complexity. It does not explain return value structure, pagination behavior for the limit parameter, or how the analysis aggregates across the specified timeperiods, leaving gaps an agent must discover through trial.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, establishing a baseline of 3. The description mentions analyzing 'top trader wallets' (plural) which conceptually maps to the walletAddress parameter, but it fails to clarify the interaction: whether providing a walletAddress analyzes that specific address versus returning a list of top traders when omitted. No additional syntax, format details, or parameter dependencies are provided beyond the schema.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool performs 'flow analysis' and calculates 'performance metrics' on 'top trader wallets,' providing specific verbs and resources. However, it fails to differentiate from the sibling tool 'hyperliquid_smart_money' or explain what distinguishes this general smart money analysis from other wallet analysis tools like 'whale_tracking' or 'copy_trading_intel'.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides no guidance on when to use this tool versus the numerous sibling alternatives (e.g., 'hyperliquid_smart_money', 'wallet_portfolio', 'copy_trading_leaderboard'). It does not clarify prerequisites, such as whether a specific walletAddress is required or if the tool returns a list of top traders when called without parameters.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

strategy_backtestBInspect

Trading Strategy Backtest: Historical backtesting simulations with Sharpe ratio, max drawdown, win rate, and P&L metrics.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full burden. It discloses output metrics (Sharpe ratio, etc.) which helps, but fails to clarify whether this tool performs live computations or retrieves cached results, execution time expectations, or whether all parameters are truly optional given the empty required array.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description efficiently packs specific financial metrics into a single sentence. The 'Trading Strategy Backtest:' prefix effectively substitutes for the null title. No redundant phrases, though it could specify the retrieval vs. computation nature of the tool.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Critical gap: with 0 required parameters and no output schema, the description fails to explain how the trading strategy itself is specified (missing parameter? predefined strategies?). It also doesn't describe default behavior when called with no arguments, which is essential for a tool with all-optional parameters.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, establishing a baseline of 3. The description implies that 'limit' controls result sets containing the mentioned metrics but does not add syntax details, validation rules, or semantic relationships between parameters (e.g., whether timeperiod affects data granularity).

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states it performs 'Historical backtesting simulations' and specifies exact output metrics (Sharpe ratio, max drawdown, win rate, P&L), which distinguishes it from generic analysis tools. However, it doesn't differentiate from sibling tools like 'einstein_meta_strategy' or 'technical_analysis'.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance is provided on when to select this tool versus alternatives like 'technical_analysis', 'general_analysis', or 'einstein_meta_strategy'. The description only states what it does, not when to use it or prerequisites for effective use.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

technical_analysisCInspect

Technical Analysis: RSI, MACD, Bollinger Bands, EMA crossover, VWAP, ATR, Volume Spike, and Stochastic RSI analysis for any token.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Since no annotations are provided, the description carries the full burden. It successfully enumerates the specific indicators calculated (RSI, VWAP, ATR, etc.), but omits critical behavioral context such as data freshness, caching behavior, rate limits, or whether the analysis is real-time versus historical.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Efficient single-sentence structure front-loaded with the indicator list. The colon format effectively categorizes the content. However, the mention of 'token' without corresponding schema support creates unnecessary ambiguity, slightly undermining the clarity.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Critical gap in explaining how to specify target tokens (given 'token' is referenced but absent from schema). With no output schema provided, the description should explain return format (e.g., whether it returns buy/sell signals, raw indicator values, or trend classifications), but it does not.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters2/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100%, establishing a baseline of 3, but the description mentions analysis 'for any token' while the input schema contains no 'token' parameter (only chain, limit, timeperiod). This creates ambiguity about how target tokens are specified. The description adds no explanatory context for the existing parameters beyond their schema definitions.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

Clearly specifies the technical indicators provided (RSI, MACD, Bollinger Bands, etc.), distinguishing it from the more generic 'general_analysis' sibling. However, it doesn't clarify the scope limitations or coverage (e.g., whether it analyzes all tokens on a chain or requires specific token selection).

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance on when to use this tool versus 'general_analysis' or other market analysis siblings like 'market_movers' or 'smart_money'. Doesn't mention that all parameters are optional or suggest default workflows.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

token_snipingBInspect

Token Sniping Analysis: Early buyers analysis with bot detection and entry timing.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
tokenAddressYesToken contract address to analyze
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description carries the full burden. It discloses the analytical behavior (bot detection, entry timing analysis) but lacks operational details such as data freshness, rate limits, authentication requirements, or whether the analysis covers historical vs real-time data. It does not describe the return format or structure.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Extremely concise single-sentence description with no redundant information. The colon structure effectively categorizes then specifies. While efficient, it borders on under-specification given the lack of annotations and output schema.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

With 100% schema coverage for 4 parameters, the input side is well handled by the schema. However, no output schema exists and the description provides no hint about return values, analysis depth, or data granularity. Adequate but incomplete for a complex analytical tool with many siblings.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, with all four parameters (tokenAddress, chain, limit, timeperiod) fully documented in the schema. The description adds no additional parameter context, but given the complete schema coverage, baseline score 3 is appropriate.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly identifies the tool's function as early buyer analysis with specific focus on bot detection and entry timing. While it uses the tool name in the opening ('Token Sniping Analysis'), it expands with specific analytical capabilities (bot detection, entry timing) that distinguish it from sibling tools like smart_money or whale_tracking.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance provided on when to use this tool versus alternatives. Given the extensive sibling list including copy_trading_intel, smart_money, and holder_concentration, the description fails to specify scenarios where bot detection and sniping analysis are preferred over general holder analysis or copy trading intelligence.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

top_gainersCInspect

Top Gainers Analysis: Top gainers/losers analysis with volume, price change %, and market insights.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full burden. It partially discloses behavioral traits by mentioning the analysis includes 'volume, price change %, and market insights', hinting at return content. However, it lacks critical safety information (read-only vs destructive), caching behavior, or data freshness guarantees.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness3/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is brief (one sentence) but structurally inefficient due to redundancy: 'Top Gainers Analysis: Top gainers/losers analysis' repeats the concept. The colon usage creates a title-like prefix that wastes space. The content after the colon is substantive but could be more direct: 'Analyze top gainers and losers with...'

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

With no output schema, the description should ideally describe return values; it vaguely mentions 'market insights' but doesn't clarify data structure or format. No annotations means safety profile is undocumented. Parameter coverage is complete via schema, making this adequate but not comprehensive for a data retrieval tool with 3 optional parameters.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Input schema has 100% description coverage with clear enum values and defaults, establishing a baseline of 3. The description mentions 'volume, price change %' which provides context for what the 'timeperiod' parameter affects, but does not add specific semantic guidance on parameter interactions or selection criteria beyond the schema.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose3/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description states the tool performs analysis of top gainers/losers with specific metrics (volume, price change %, market insights). However, it creates confusion by mentioning 'losers' when the tool name is 'top_gainers', and fails to distinguish from sibling 'market_movers'. The opening 'Top Gainers Analysis:' restates the tool's purpose tautologically before adding details.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance provided on when to use this tool versus siblings like 'market_movers', 'general_analysis', or 'technical_analysis'. No prerequisites, filtering recommendations, or exclusions are mentioned despite the tool having 0 required parameters.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

wallet_portfolioCInspect

Wallet Portfolio Analysis: Complete wallet portfolio breakdown with holdings and performance.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
walletAddressNoWallet address to analyze
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full burden of behavioral disclosure but offers minimal detail. While it implies a read-only analysis operation, it omits critical context such as data freshness, caching behavior, rate limits, or what happens when analyzing wallets with extensive holdings (despite the presence of a 'limit' parameter).

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness3/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is brief but suffers from redundancy ('Wallet Portfolio Analysis' followed by 'Complete wallet portfolio breakdown'). The colon structure front-loads the category but wastes space restating the tool name rather than adding actionable guidance.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the absence of annotations and output schema, the description is insufficient for a 4-parameter tool with optional filtering across 7 chains and 5 time periods. It fails to describe the return structure, pagination behavior (despite the 'limit' parameter), or the specific metrics included in 'performance' analysis.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The input schema has 100% description coverage, establishing a baseline of 3. The description mentions 'performance' which loosely maps to the 'timeperiod' parameter, but adds no specific syntax guidance, format examples, or semantic relationships between parameters (e.g., how 'chain' filters the analysis).

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool performs a 'Complete wallet portfolio breakdown' covering 'holdings and performance,' providing specific verbs and scope. However, it does not explicitly differentiate from sibling tools like 'general_analysis' or 'wallet_trust_score' that might also analyze wallet data.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance is provided on when to use this tool versus alternatives (e.g., 'general_analysis' for broader market context or 'wallet_trust_score' for risk assessment). There are no prerequisites, warnings, or exclusion criteria mentioned.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

wallet_trust_scoreBInspect

Wallet Trust Score: Score wallets by historical accuracy, profitability, consistency, and behavior. Identifies whales, smart money, KOLs, and bots.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full burden. It successfully discloses the scoring criteria (accuracy, profitability, consistency, behavior) and classification targets, adding valuable context. However, it lacks operational details such as score ranges, data freshness, rate limits, or whether results are ranked.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two efficiently structured sentences with zero redundancy. The first sentence establishes the core scoring mechanism, the second the identification capabilities—every word earns its place. Front-loaded with the most critical information.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Adequate for a tool with 3 simple parameters and no output schema, explaining what the tool returns conceptually (scores and classifications). However, given the absence of an output schema, it should specify the return structure or score format, and it lacks differentiation guidance for overlapping sibling tools.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, establishing a baseline of 3. The description mentions 'Score wallets' but adds no semantic meaning about how parameters (chain, limit, timeperiod) affect the scoring logic or result set beyond what the schema already provides.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool scores wallets using specific metrics (historical accuracy, profitability, consistency, behavior) and identifies specific actor types (whales, smart money, KOLs, bots). However, it loses a point because it claims to identify 'whales' and 'smart money' without distinguishing from sibling tools 'whale_tracking' and 'smart_money', potentially causing selection confusion.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance provided on when to use this tool versus siblings with overlapping functionality (whale_tracking, smart_money, wallet_portfolio). No prerequisites, exclusions, or alternative recommendations are mentioned.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

whale_trackingCInspect

Whale Intelligence: Token-centric whale accumulation view with net capital flows and smart money tracking.

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries full disclosure burden. It implies read-only behavior via 'view' and hints at return data structure ('accumulation', 'net capital flows'), but omits rate limits, caching behavior, data freshness, and specific return format.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness3/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Single sentence avoids bloat, but 'Whale Intelligence:' prefix wastes space with category labeling rather than actionable information. Dense jargon ('smart money', 'net capital flows') packs information but requires domain knowledge to parse.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Insufficient for a 3-parameter blockchain analytics tool lacking output schema and annotations. Description does not clarify that all parameters are optional, what 'whale' threshold defines results, or how 'token-centric' manifests without a token input parameter.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100%, establishing baseline 3. Description mentions 'Token-centric' which adds domain context (results organized by token) despite no token filter parameter, but this slightly confuses the tool's input requirements. No additional syntax or format guidance provided.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose3/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description conveys the domain (whale/accumulation tracking, capital flows) but uses marketing language ('Whale Intelligence', 'smart money') rather than clear operational verbs. It fails to differentiate from sibling 'smart_money' or 'copy_trading_intel' tools, leaving ambiguity about unique value.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance provided on when to prefer this tool over alternatives like 'smart_money', 'holder_concentration', or 'copy_trading_intel'. Does not mention prerequisites (e.g., whether specific chains work better) or when-not-to-use conditions.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

zora_launchesCInspect

Zora Launchpad: Zora Launchpad token launches and volume analysis (Base only).

ParametersJSON Schema
NameRequiredDescriptionDefault
chainNoBlockchain network to query
limitNoMaximum number of results (default: 10, max: 100)
timeperiodNoTime period for analysis
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are present, so the description carries full disclosure burden. It provides one behavioral constraint ('Base only'), but this directly contradicts the input schema which accepts multiple chain values (ethereum, bsc, solana, etc.). No information on return format, pagination, rate limits, or mutation safety is provided.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness3/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is brief (one sentence) but inefficiently structured with redundant phrasing ('Zora Launchpad: Zora Launchpad...'). The parenthetical '(Base only)' is appropriately placed at the end, but the colon-separated prefix restates the tool name without adding value.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

With no output schema and no annotations, the description should explain what data is returned (e.g., launch metrics, volume calculations). It fails to resolve the ambiguity between the 'Base only' claim and the multi-chain schema, leaving agents uncertain about valid chain inputs. Essential behavioral context for a financial data tool is missing.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, establishing a baseline of 3. The description attempts to add semantic constraint with 'Base only' regarding the chain parameter, though this conflicts with the schema's enum values. No additional context is provided for timeperiod granularity or limit behavior beyond the schema descriptions.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly identifies the resource (Zora Launchpad tokens) and actions (launches and volume analysis). It specifies 'Base only' to scope the tool's domain. However, it redundantly repeats 'Zora Launchpad' and does not explicitly distinguish from the sibling 'pump_fun_launches' tool.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance provided on when to use this tool versus alternatives like 'pump_fun_launches' or 'general_analysis'. No prerequisites, exclusions, or workflow context is mentioned. The description only states what the tool does, not when to invoke it.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

Discussions

No comments yet. Be the first to start the discussion!

Try in Browser

Your Connectors

Sign in to create a connector for this server.

Resources