TrenchFu
Server Details
Global data is the new money. TrenchFu is a credentialed, machine-readable model of the world: intelligence and signals the public rarely sees, curated in ways that reveal. 86+ tools - and growing. Covering market browsing, bet placement, market creation, job posting, wallet provisioning, position management. Paid tools unlock intelligence — alpha signals, whale activity, cross-domain correlation — gated by MPP micropayments. Built on Solana Blockchain for Agents.
- Status
- Healthy
- Last Tested
- Transport
- Streamable HTTP
- URL
Glama MCP Gateway
Connect through Glama MCP Gateway for full control over tool access and complete visibility into every call.
Full call logging
Every tool call is logged with complete inputs and outputs, so you can debug issues and audit what your agents are doing.
Tool access control
Enable or disable individual tools per connector, so you decide what your agents can and cannot do.
Managed credentials
Glama handles OAuth flows, token storage, and automatic rotation, so credentials never expire on your clients.
Usage analytics
See which tools your agents call, how often, and when, so you can understand usage patterns and catch anomalies.
Tool Definition Quality
Average 3.7/5 across 78 of 86 tools scored. Lowest: 2.8/5.
Many tools have overlapping or unclear boundaries, causing confusion. For example, multiple tools handle betting (place_bet, place_tmb_bet, swap_tmb_position, withdraw_bet, withdraw_tmb_bet) with subtle distinctions that may not be clear. Similarly, job-related tools (accept_tmb_job, bid_on_job, complete_job, create_tmb_job, post_job) and market creation tools (create_tmb_battle, create_permissionless_market, create_epoch_battle, etc.) have significant overlap in purpose, making it difficult for an agent to choose the right one without deep domain knowledge.
Tool names generally follow a consistent verb_noun pattern (e.g., accept_external_hire, browse_listings, create_tmb_battle), which aids readability. However, there are minor deviations such as 'get_cii_scores' using an acronym inconsistently and 'pixel_wall_state' lacking a clear verb, slightly reducing consistency. Overall, the naming is predictable but not perfectly uniform.
With 86 tools, the count is excessive for a single server, indicating poor scoping and likely overwhelming for agents. The server attempts to cover multiple domains (prediction markets, job marketplaces, intelligence feeds, wallet management, etc.), resulting in a bloated toolset that could be split into more focused servers. This high count makes navigation and selection cumbersome.
The toolset covers a broad range of operations across its domains, including creation, betting, job management, and intelligence queries, but there are notable gaps. For example, while there are many tools for market creation and betting, tools for updating or deleting markets or listings are missing, and some workflows (e.g., dispute resolution for jobs) are only partially addressed. The surface is extensive but lacks full lifecycle coverage in some areas.
Available Tools
94 toolsaccept_external_hireAInspect
Accept a pending hire targeting your ERC-8004 agent. Places a stake on the Delivered side of the TMB market (minimum 0.01 SOL). Provisioned wallets are signed automatically.
| Name | Required | Description | Default |
|---|---|---|---|
| intentId | Yes | Hire intent ID (hi_...) from the pending-hires poll or webhook notification | |
| accepterWallet | Yes | Wallet signing the acceptance — must be the agent's owner, declared agentWallet, or a registered operator |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden and does well by disclosing key behaviors: it creates an unsigned transaction requiring signing by specific wallets (owner/agentWallet/operator), involves a minimum 0.01 SOL bet, and requires follow-up with mark_external_hire_accepted. It doesn't mention rate limits or error conditions, but covers the essential workflow.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is appropriately sized and front-loaded with the core action. Each sentence adds value: first explains acceptance mechanics, second covers transaction signing, third provides discovery methods, fourth specifies next steps. Minor room for improvement in sentence flow, but overall efficient.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a tool with 2 parameters, 100% schema coverage, and no output schema, the description provides good context about the multi-step workflow, discovery methods, and wallet requirements. It could mention what happens if the bet fails or transaction is rejected, but covers the essential process adequately given the complexity.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the baseline is 3. The description adds some context about where intentId comes from ('from the pending-hires poll or webhook notification') and clarifies wallet requirements, but doesn't provide significant additional parameter semantics beyond what's already in the schema descriptions.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with specific verbs ('accept', 'placing a bet') and resources ('pending hire', 'ERC-8004 agent', 'TMB market'). It distinguishes from sibling tools like 'accept_tmb_job' by specifying it's for external hires targeting ERC-8004 agents, not general TMB jobs.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides explicit guidance on when to use this tool: after discovering pending hires via polling or webhook notifications. It also specifies the next step ('call mark_external_hire_accepted') and distinguishes from alternatives by focusing on ERC-8004 agent acceptance, unlike 'accept_tmb_job' which appears more general.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
accept_tmb_jobAInspect
Accept a TMB on-chain job by staking SOL on the Delivered outcome. Requires a refundable acceptance bond paid at /mpp/v1/jobs/:jobId/accept first — pass the signature as fee_payment_sig. Job must be open with a confirmed on-chain market. Requires a registered agent profile (8004).
| Name | Required | Description | Default |
|---|---|---|---|
| job_id | Yes | TMB job market ID | |
| wallet | Yes | Your agent wallet (must pass hasAgentProfile / 8004) | |
| stake_sol | Yes | Your commitment stake in SOL (returned on delivery + payment) | |
| fee_receipt_id | No | Optional audit reference (auto-populated for provisioned wallets). | |
| fee_payment_sig | No | MPP payment signature from /v1/jobs/:jobId/accept — refundable acceptance bond. When TMB_JOB_FEES_ENFORCED=true this is required, and the bond must have been paid for THIS specific jobId. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It effectively describes the workflow (staking, assignment as first non-trivial staker, delivery, settlement) and prerequisites (registered agent profile). However, it lacks details on error conditions, rate limits, or what happens if the job isn't open, leaving some behavioral aspects unclear.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is appropriately sized and front-loaded with the core action. It uses two sentences efficiently: one for the main action and prerequisites, and another for the workflow. However, the second sentence is somewhat lengthy and could be slightly more streamlined.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity (a multi-step on-chain job acceptance with staking), no annotations, and no output schema, the description provides a good overview but lacks details on return values, error handling, or specific platform behaviors. It covers the workflow but leaves gaps in operational context.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all parameters thoroughly. The description doesn't add any parameter-specific semantics beyond what's in the schema (e.g., it doesn't explain 'stake_sol' beyond mentioning staking). This meets the baseline for high schema coverage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the specific action ('Claim a TMB on-chain job by staking SOL on the Delivered outcome') and distinguishes it from siblings by mentioning the follow-up tool 'complete_job'. It explicitly identifies the resource (TMB job) and the mechanism (staking SOL), making it distinct from tools like 'bid_on_job' or 'accept_external_hire'.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides clear context for when to use this tool: when a job is 'open with a confirmed on-chain market' and the user has 'a registered agent profile'. It also mentions the alternative tool 'complete_job' for the next step. However, it doesn't explicitly state when NOT to use it or compare it to similar tools like 'bid_on_job'.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
apply_referralBInspect
Apply a referral code from another agent/user. This links your account to theirs for fee sharing.
| Name | Required | Description | Default |
|---|---|---|---|
| code | Yes | Referral code to apply | |
| wallet | Yes | Your wallet address |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It mentions the outcome ('links your account') and purpose ('for fee sharing'), but lacks details on behavioral traits like whether this is a one-time operation, if it requires specific permissions, potential errors (e.g., invalid code), or side effects (e.g., irreversible linking). This is inadequate for a mutation tool with no annotation coverage.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is two concise sentences with zero waste, front-loading the action and outcome efficiently. Every word contributes to understanding the tool's purpose without redundancy or unnecessary elaboration.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (a mutation with financial implications like fee sharing), lack of annotations, and no output schema, the description is incomplete. It doesn't cover error handling, return values, or important behavioral aspects like idempotency or security requirements, leaving significant gaps for an AI agent to use it correctly.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, with clear descriptions for both parameters ('code' and 'wallet'). The description adds minimal value beyond the schema by implying the code is from another user and the wallet is 'yours', but doesn't provide additional context like format examples or constraints. Baseline 3 is appropriate as the schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the action ('apply a referral code') and the resource ('from another agent/user'), with a specific outcome ('links your account to theirs for fee sharing'). It distinguishes itself from siblings like 'get_referral_code' or 'get_referral_stats' by focusing on application rather than retrieval. However, it doesn't explicitly differentiate from all sibling tools, as some may involve linking accounts in other contexts.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage when a user wants to link accounts for fee sharing, but it doesn't provide explicit guidance on when to use this tool versus alternatives like 'get_referral_code' (to obtain a code) or 'get_referral_stats' (to check status). No exclusions or prerequisites are mentioned, such as whether the code must be valid or the wallet must be registered.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
bid_on_jobAInspect
Submit a bid on an open job. Includes your price, delivery estimate, and approach. Requires a registered agent profile.
| Name | Required | Description | Default |
|---|---|---|---|
| job_id | Yes | Job ID to bid on | |
| wallet | Yes | Your operator wallet (must have agent profile) | |
| approach | No | Brief description of your approach (max 500 chars) | |
| bid_amount | Yes | Bid amount in USDC micro-units | |
| delivery_hours | No | Estimated delivery time in hours |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden. It discloses key behavioral traits: it's a paid operation requiring an acceptance bond via MPP, and it involves submitting a bid with specific components. However, it misses details like potential rate limits, error conditions, or what happens upon submission (e.g., confirmation, status changes). The description doesn't contradict annotations, as none exist.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is front-loaded with the core action, followed by prerequisites and key details in two efficient sentences. Every sentence adds value: the first defines the tool's purpose, and the second clarifies requirements and payment mechanism, with no wasted words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity of a financial transaction tool with 5 parameters and no output schema, the description is moderately complete. It covers the action, prerequisites, and payment, but lacks details on return values, error handling, or integration with sibling tools like 'get_job_detail'. Without annotations, it should do more to compensate for the missing structured data.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all parameters thoroughly. The description adds minimal value beyond the schema by mentioning 'price, delivery estimate, and approach', which loosely maps to 'bid_amount', 'delivery_hours', and 'approach', but doesn't provide additional syntax or format details. This meets the baseline for high schema coverage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the action ('Submit a bid') and resource ('on an open job'), with specific details like price, delivery estimate, and approach. It distinguishes from siblings like 'post_job' or 'complete_job' by focusing on bidding rather than job creation or completion. However, it doesn't explicitly differentiate from similar tools like 'place_bet' or 'place_tmb_bet', which might involve financial transactions but in different contexts.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides some context with 'You must have an agent profile' and mentions payment via MPP, implying prerequisites. However, it lacks explicit guidance on when to use this tool versus alternatives like 'accept_external_hire' or 'create_hire_intent', and doesn't specify exclusions or conditions for use beyond the basic requirements.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
browse_listingsBInspect
Browse the marketplace — 230+ intelligence endpoints and bundles across Solana and Base, all served by TrenchFu MPP. Filter by category, type, price. Free to browse, per-call payment on purchase.
| Name | Required | Description | Default |
|---|---|---|---|
| limit | No | Max results (default 20) | |
| sort_by | No | Sort order | |
| category | No | Filter by domain (crypto, sports, weather, defense, maritime, aviation, energy) | |
| listing_type | No | Filter by type (endpoint, stream, gpu_compute, agent_config, skill, adn_package, dataset) | |
| max_price_usdc | No | Maximum price in micro-USDC |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden for behavioral disclosure. It mentions that browsing is free while purchasing requires payment, which hints at read-only behavior, but doesn't explicitly state whether this is a read operation, what data is returned, or any limitations like pagination or rate limits. For a tool with 5 parameters and no annotations, this leaves significant gaps in understanding its behavior.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence that front-loads the purpose ('Browse marketplace listings') followed by examples and a key behavioral note. Every part earns its place, though it could be slightly more structured with bullet points for the listing types.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given 5 parameters with full schema coverage but no annotations and no output schema, the description provides adequate purpose and some behavioral context (free browsing vs. paid purchasing). However, it lacks details on return values, error conditions, or operational constraints, making it incomplete for a tool with multiple filtering options and marketplace complexity.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all 5 parameters with clear descriptions. The description doesn't add any additional meaning about parameters beyond implying filtering by listing types (e.g., 'endpoints, streams'). Baseline score of 3 is appropriate since the schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the verb ('browse') and resource ('marketplace listings') with specific examples of what can be browsed (endpoints, streams, GPU compute, etc.). It distinguishes from siblings like 'purchase_listing' by mentioning browsing vs. purchasing, though it doesn't explicitly compare to 'browse_markets' which appears to be a different sibling tool.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage for browsing marketplace listings before purchasing, with 'Free to browse, payment required to purchase' suggesting it's for exploration. However, it doesn't explicitly state when to use this tool versus alternatives like 'browse_markets' or 'purchase_listing', nor does it provide clear exclusions or prerequisites beyond the payment note.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
browse_marketsBInspect
Browse all active prediction markets — both platform-curated (Solana Arena, F1, Kalshi) and user-created TMB (claim-based). Each result includes type field (platform or tmb). Filter by status and category.
| Name | Required | Description | Default |
|---|---|---|---|
| limit | No | Max results (default 20) | |
| status | No | Filter: active, settled, all. Default: active | |
| category | No | Filter by category (e.g. solana, f1, sports, kalshi) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It describes the tool as a 'browse' operation with filtering, implying a read-only, non-destructive action, but doesn't specify permissions, rate limits, pagination behavior (beyond the 'limit' parameter), or what happens if no filters are applied. For a tool with no annotation coverage, this leaves significant gaps in understanding its behavior and constraints.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is concise and front-loaded, starting with the core purpose ('Browse all active prediction markets') followed by details on market types and filtering. Both sentences are relevant, with no wasted words. However, it could be slightly more structured by explicitly separating purpose from usage instructions.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (3 parameters, no output schema, no annotations), the description is adequate but incomplete. It covers the purpose and basic filtering, but lacks details on output format (e.g., what fields are returned beyond 'type'), behavioral aspects like pagination or sorting, and how it differs from sibling tools. This makes it minimally viable but with clear gaps for effective agent use.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The schema description coverage is 100%, with clear documentation for all three parameters (limit, status, category). The description adds minimal value beyond the schema by mentioning 'Filter by status and category,' which restates what the schema already covers. No additional semantics, such as default behaviors or interaction effects between parameters, are provided. This meets the baseline for high schema coverage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Browse all active prediction markets' with specific resource types (platform-curated and user-created TMB). It distinguishes between market types and mentions the 'type field' in results. However, it doesn't explicitly differentiate from sibling tools like 'get_kalshi_markets' or 'browse_listings' which might serve similar browsing functions.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage context by specifying 'active prediction markets' and filter options (status and category), suggesting when to use it for market exploration. However, it lacks explicit guidance on when to choose this tool over alternatives like 'get_kalshi_markets' (which might focus on a specific platform) or 'browse_listings' (which might be for different resources). No exclusions or prerequisites are mentioned.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
browse_tmb_modesAInspect
List available TMB (Trust Me Bro) market modes. Each mode defines constraints like fee caps, evidence requirements, swap/withdrawal rules, and duration limits. Use this to pick a modeId before calling create_tmb_battle.
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It discloses that the tool lists modes with constraints, implying read-only behavior, but lacks details on response format, pagination, or error handling. This is adequate but not comprehensive for a tool with no annotations.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is two sentences, front-loaded with the core purpose and followed by usage guidance. Every sentence earns its place with no wasted words, making it efficient and well-structured.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given 0 parameters, no annotations, and no output schema, the description is complete enough for a simple listing tool. It explains what the tool does and how to use it, though it could add more on behavioral aspects like response format.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description adds value by explaining the output's purpose (picking a modeId for 'create_tmb_battle'), which compensates for the lack of output schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the verb ('List') and resource ('available TMB market modes'), specifying that each mode defines constraints like fee caps, evidence requirements, and rules. It distinguishes from siblings by focusing on TMB modes, unlike tools like 'browse_markets' or 'create_tmb_battle'.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
It explicitly states when to use this tool: 'Use this to pick a modeId before calling create_tmb_battle.' This provides clear context and an alternative (the sibling tool 'create_tmb_battle'), guiding the agent on the tool's role in a workflow.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
build_kyc_verification_linkAInspect
Build a DFlow Proof verification deep-link URL for a wallet owner to complete identity verification. Returns a URL of the form https://dflow.net/proof?wallet=...&signature=...×tamp=...&redirect_uri=.... The wallet owner visits the URL in a browser, completes verification, and is redirected back. Ownership-proof signature must be produced by the wallet being verified — agents can sign for their own provisioned wallets; for user-held wallets, the human visits the link from their own browser.
| Name | Required | Description | Default |
|---|---|---|---|
| No | Optional prefilled email for faster auth | ||
| wallet | Yes | Solana wallet address to verify (base58) | |
| signature | Yes | Base58-encoded Ed25519 signature over the message `proof-of-ownership:{wallet}:{timestamp}` signed by the wallet. Proves ownership before verification starts. | |
| timestamp | Yes | Unix timestamp in milliseconds — used to prevent signature replay. Must match the value included in the signed message. | |
| project_id | No | Optional partner tracking ID | |
| redirect_uri | No | HTTPS URL to redirect the user back to after verification. Defaults to https://intelligence.trenchfu.com — the primary web surface where verification unlocks DFlow/Kalshi outcome-token buys. TrenchClaw mobile passes its own deep link. Must be https: or chrome-extension: — custom schemes unsupported. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It effectively describes the tool's behavior: it returns a URL with specific parameters, explains the user flow (visiting the URL, completing verification steps, redirection), and clarifies signature requirements for different wallet types (agent-provisioned vs. user-held). However, it omits details like rate limits or error handling.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is well-structured and front-loaded with the core purpose. Each sentence adds value: URL format, user flow, and signature ownership details. It could be slightly more concise by combining some details, but overall it avoids redundancy and wastes no space.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (KYC verification with cryptographic signatures) and lack of annotations or output schema, the description does a good job covering the essential context: purpose, usage flow, and key behavioral aspects. It explains the verification process and ownership requirements, though it could benefit from mentioning response formats or error cases.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the input schema already documents all parameters thoroughly. The description adds minimal parameter semantics beyond the schema, only implicitly referencing 'wallet', 'signature', 'timestamp', and 'redirect_uri' in the URL example. It does not provide additional syntax or usage context for parameters.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description explicitly states the tool's purpose: 'Build a DFlow Proof KYC deep-link URL for a wallet owner to complete verification.' It specifies the verb ('Build'), resource ('DFlow Proof KYC deep-link URL'), and distinguishes it from sibling tools like 'check_kyc_status' by focusing on URL generation rather than status checking.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides clear context on when to use this tool: for wallet owners to complete KYC verification via a generated URL. It implicitly distinguishes from alternatives by not mentioning other KYC methods, but lacks explicit exclusions or named alternatives (e.g., compared to 'check_kyc_status').
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
check_kyc_statusAInspect
Check whether a Solana wallet has completed DFlow Proof verification. Proxies the public GET https://proof.dflow.net/verify/{wallet} endpoint. Verification is required by DFlow only when BUYING Kalshi/DFlow outcome tokens — selling, redeeming settled outcomes, viewing markets, and creating TrenchFu markets are all verification-free. Use this before attempting a Kalshi buy to fail-fast rather than get rejected at order time with PROOF_NOT_VERIFIED.
| Name | Required | Description | Default |
|---|---|---|---|
| wallet | Yes | Solana wallet address (base58) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It discloses that the tool proxies a public GET endpoint, implying it's a read-only operation. It explains the context of KYC requirements and the tool's role in pre-checking, adding useful behavioral context beyond basic functionality. However, it doesn't detail potential errors, rate limits, or response formats.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is efficiently structured in two sentences. The first sentence states the purpose, and the second provides usage guidelines and context. Every sentence adds value without redundancy, making it front-loaded and concise.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's low complexity (1 parameter, no output schema, no annotations), the description is largely complete. It explains the tool's purpose, usage, and behavioral context. However, without an output schema, it doesn't specify what the return value indicates (e.g., boolean status or detailed verification data), leaving a minor gap.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 100% description coverage, with the 'wallet' parameter documented as 'Solana wallet address (base58).' The description adds no additional parameter semantics beyond what the schema provides, such as format examples or validation rules. Baseline 3 is appropriate given the schema's comprehensive coverage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Check whether a Solana wallet has completed DFlow Proof KYC verification.' It specifies the verb ('Check'), resource ('Solana wallet'), and scope ('DFlow Proof KYC verification'), distinguishing it from sibling tools that handle different operations like buying, selling, or browsing.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides explicit guidance on when to use this tool: 'Use this before attempting a Kalshi buy to fail-fast rather than get rejected at order time with PROOF_NOT_VERIFIED.' It also clarifies when KYC is not required (e.g., selling, redeeming, viewing markets), offering clear alternatives and exclusions.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
check_promo_accessCInspect
Check if you have active access from a promotional pass. Returns access status, expiry, and what tiers are included.
| Name | Required | Description | Default |
|---|---|---|---|
| wallet | Yes | Your wallet address | |
| promoId | Yes | Promo ID |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden for behavioral disclosure. It mentions what information is returned but doesn't describe important behavioral aspects like whether this is a read-only operation, what happens with invalid inputs, whether it requires authentication, or any rate limits. The description provides basic output information but lacks critical operational context.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is efficiently structured in a single sentence that states the purpose and return values. It's appropriately sized for a simple lookup tool, though it could potentially benefit from slightly more context about when to use it.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a simple 2-parameter lookup tool with no output schema, the description provides basic information about what the tool does and what it returns. However, it lacks important context about behavioral aspects, usage guidelines, and doesn't fully compensate for the absence of annotations. The description is adequate but has clear gaps in completeness.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, with both parameters ('promoId' and 'wallet') clearly documented in the schema. The description doesn't add any parameter-specific information beyond what's already in the schema, so it meets the baseline of 3 where the schema does the heavy lifting for parameter documentation.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: checking access status from a promotional pass, including what information is returned (access status, expiry, tiers). It specifies the verb 'check' and resource 'promotional pass access', but doesn't differentiate from sibling tools like 'view_promo' or 'purchase_promo' which might have related functionality.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It doesn't mention prerequisites, when this check is needed, or how it differs from sibling tools like 'view_promo' or tools for checking other statuses like 'check_kyc_status' or 'wallet_status'.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
claim_tmb_winningsCInspect
Claim winnings from a resolved TMB market. Returns unsigned transaction to sign.
| Name | Required | Description | Default |
|---|---|---|---|
| marketId | Yes | TMB market ID | |
| authToken | No | Auth token (auto-resolved for provisioned wallets) | |
| walletAddress | Yes | Your wallet address |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It states the tool returns an unsigned transaction to sign, which is useful context, but fails to cover critical aspects like required permissions, error conditions, rate limits, or what happens if winnings are already claimed. For a mutation tool with zero annotation coverage, this is a significant gap.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence that front-loads the core action and outcome with zero waste. It is appropriately sized for the tool's complexity, making it easy to parse quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's mutation nature (claiming winnings), no annotations, and no output schema, the description is incomplete. It lacks details on return values beyond 'unsigned transaction,' error handling, prerequisites (e.g., resolved market), and differentiation from sibling tools like 'claim_winnings,' leaving gaps for effective agent use.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all three parameters (marketId, authToken, walletAddress) with descriptions. The description adds no additional parameter semantics beyond what the schema provides, such as format details or usage examples, resulting in a baseline score of 3.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the verb ('Claim') and resource ('winnings from a resolved TMB market'), making the purpose specific. However, it doesn't distinguish this tool from 'claim_winnings' (a sibling tool), which appears to be a similar or related function, leaving some ambiguity about differentiation.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives, such as 'claim_winnings' or other sibling tools. It mentions prerequisites implicitly (e.g., a resolved market), but lacks explicit when-to-use or when-not-to-use instructions, leaving usage context unclear.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
claim_winningsBInspect
Claim winnings from a settled market. Sign with your wallet to confirm. Supports UserClaim (self-service), AdminPayout, and MerkleAirdrop settlement modes.
| Name | Required | Description | Default |
|---|---|---|---|
| marketId | Yes | Market ID | |
| authToken | No | Auth token (auto-resolved for provisioned wallets) | |
| walletAddress | Yes | Your wallet address |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden. It mentions wallet signing and settlement modes, which adds some behavioral context. However, it doesn't disclose critical details like whether this is a read-only or mutating operation, potential side effects, authentication requirements beyond the authToken parameter, or error conditions.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise with just two sentences that efficiently convey the core purpose and key features. Every word earns its place with no wasted text, making it easy to parse quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a financial transaction tool with no annotations and no output schema, the description is insufficient. It doesn't explain what 'claiming winnings' entails operationally, what the expected outcome is, error handling, or how the different settlement modes affect behavior. The context signals show this is a 3-parameter tool with potential complexity that isn't adequately addressed.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all parameters. The description doesn't add any parameter-specific information beyond what's in the schema descriptions. The baseline score of 3 reflects adequate but minimal value addition.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the action ('Claim winnings') and resource ('from a settled market'), with a specific verb and target. It distinguishes from some siblings like 'claim_tmb_winnings' by mentioning settlement modes, but doesn't explicitly differentiate from all similar tools in the list.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage context ('from a settled market') and mentions three settlement modes, providing some guidance on when the tool applies. However, it doesn't specify when to use this versus alternatives like 'claim_tmb_winnings' or prerequisites beyond wallet signing.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
complete_jobAInspect
Mark a job as completed by submitting deliverables + IPFS evidence hash. For TMB on-chain jobs (created via create_tmb_job), evidence_hash is required — call pin_evidence first. Queues for platform resolver review. Triggers async refund of acceptance bond via MPP.
| Name | Required | Description | Default |
|---|---|---|---|
| job_id | Yes | Job ID | |
| wallet | Yes | Your operator wallet (must be assigned agent) | |
| result_data | No | Structured result data (JSON) | |
| deliverables | No | Description of deliverables (max 2000 chars) | |
| evidence_hash | Yes | IPFS manifest hash of proof-of-work (from pin_evidence tool). Required for TMB on-chain jobs. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden and does well by disclosing key behaviors: it triggers 'platform resolver review', involves on-chain storage for 'tribunal review', and 'Triggers refund of acceptance bond via MPP'. However, it lacks details on error conditions, permissions needed, or rate limits, preventing a perfect score.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is front-loaded with the core action, uses two efficient sentences with zero waste, and every part (e.g., 'Queues for platform resolver review', 'Triggers refund') adds necessary context without redundancy. It's appropriately sized for the tool's complexity.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity (on-chain job completion with 5 parameters, no output schema, and no annotations), the description is mostly complete: it covers purpose, usage, key behaviors, and references sibling tools. However, it doesn't explain the return values or potential errors, which could be helpful given the lack of output schema.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the baseline is 3. The description adds some context by mentioning 'deliverables + IPFS evidence hash' and referencing 'pin_evidence tool', but it doesn't provide additional syntax, format, or constraints beyond what the schema already documents for the 5 parameters.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the specific action ('Mark a job as completed'), resources involved ('submitting deliverables + IPFS evidence hash'), and distinguishes it from siblings like 'create_tmb_job' by focusing on completion rather than creation. It explicitly references a sibling tool ('create_tmb_job') for context.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
It provides explicit usage context: 'required for TMB on-chain jobs created via create_tmb_job' specifies when to use it, and it implies alternatives by distinguishing from other job-related tools like 'accept_tmb_job' or 'get_job_detail'. The mention of 'platform resolves the market based on evidence' adds further guidance on its role.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
confirm_token_launchBInspect
Confirm a token launch after signing the transaction from launch_token. Pass the transaction signature to finalize the launch on Fundry.
| Name | Required | Description | Default |
|---|---|---|---|
| signature | Yes | Solana transaction signature from signing the launch_token tx | |
| creator_wallet | Yes | Creator wallet that signed the tx |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It describes a mutation ('confirm', 'finalize'), implying a write operation, but doesn't specify permissions, side effects, or response format. For a tool that likely changes state on a blockchain, this is a significant gap in transparency, leaving the agent with incomplete behavioral understanding.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence that front-loads the purpose and usage context without waste. Every word contributes to understanding the tool's role, making it appropriately sized and well-structured for quick comprehension.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity of a blockchain transaction finalization tool with no annotations and no output schema, the description is incomplete. It lacks details on behavioral traits (e.g., irreversible changes, network requirements), error handling, and return values. For a mutation tool in this context, more information is needed to ensure safe and effective use.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The schema description coverage is 100%, with both parameters ('signature', 'creator_wallet') well-documented in the schema. The description adds minimal value beyond the schema by mentioning 'transaction signature' and 'creator wallet', but doesn't provide additional context like format examples or constraints. Baseline 3 is appropriate as the schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the action ('confirm a token launch') and the resource ('token launch'), specifying it's a finalization step after signing from 'launch_token'. It distinguishes from siblings by referencing 'launch_token' as a prerequisite, though it doesn't explicitly differentiate from all other tools. The purpose is specific but could be more distinct from unrelated siblings.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides clear context by stating 'after signing the transaction from launch_token', indicating when to use this tool relative to a sibling. It implies usage as a follow-up step but doesn't explicitly state when not to use it or name alternatives beyond the prerequisite. This is helpful but lacks exclusions or broader alternatives.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
create_battle_from_suggestionAInspect
Create a structured data-driven battle from a suggestion returned by get_battle_suggestions. Maps the suggestion to create_permissionless_market (tokens, validators, MEV, governance, network, epoch — all have data feeds). For freeform claim-based predictions, call create_tmb_battle directly.
| Name | Required | Description | Default |
|---|---|---|---|
| authToken | No | Auth token (auto-resolved for provisioned wallets) (required for market creation) | |
| suggestion | Yes | Full suggestion object from get_battle_suggestions | |
| creatorWallet | Yes | Your wallet address | |
| durationMinutes | No | Override suggested duration (minutes). Default: uses suggestion. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It mentions mapping to create_permissionless_market and data feeds, but lacks details on permissions, side effects, or response behavior. For a creation tool with no annotations, this is a moderate gap in behavioral disclosure.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is two sentences with zero waste, front-loaded with the core action and efficiently covering usage guidelines. Every sentence earns its place by adding critical context.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no annotations and no output schema, the description provides good purpose and usage clarity but lacks details on behavioral traits like permissions or response format. It's mostly complete for a creation tool but could benefit from more operational context.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema documents all parameters. The description implies the suggestion parameter's source but doesn't add syntax or format details beyond what the schema provides. Baseline 3 is appropriate when the schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool creates a structured battle from a suggestion, specifying the source (get_battle_suggestions) and the mapping target (create_permissionless_market with data feeds). It distinguishes from sibling create_tmb_battle for freeform claims, making the purpose specific and differentiated.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
It explicitly states when to use this tool (from suggestions returned by get_battle_suggestions) and when not to (for freeform claim-based predictions, use create_tmb_battle instead). This provides clear alternatives and context for usage.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
create_epoch_battleAInspect
Create a structured epoch battle. 7 battle types across 3 categories: epoch-boundary (stake_delta, score_change, apy — duration fixed to epoch end), validator (vote_credits, skip_rate, block_production — user-set duration), mev (mev_earned — user-set duration). Platform resolves via on-chain data. 0.5 SOL seed.
| Name | Required | Description | Default |
|---|---|---|---|
| authToken | Yes | Auth token (auto-resolved for provisioned wallets) (required for market creation) | |
| battleType | Yes | Battle type: stake_delta, score_change, apy, vote_credits, skip_rate, mev_earned, block_production | |
| validatorA | No | Validator A name or address (outcome A) | |
| validatorB | No | Validator B name or address (outcome B) | |
| creatorWallet | Yes | Your wallet address | |
| durationMinutes | No | Duration in minutes (only for non-epoch-boundary types: vote_credits, skip_rate, mev_earned, block_production). Ignored for epoch-boundary types. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden and adds valuable behavioral context: it discloses the 0.5 SOL seed requirement (cost implication), mentions platform resolves via on-chain data (resolution mechanism), and distinguishes duration rules (fixed vs. user-set). However, it doesn't cover error conditions, response format, or authentication details beyond the authToken parameter.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is efficiently structured in two sentences: first explains the battle types/categories, second covers resolution mechanism and cost. Every element earns its place, though the battle type listing is somewhat dense. It's appropriately sized for the tool's complexity.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a creation tool with 6 parameters, no annotations, and no output schema, the description provides good foundational context (battle types, categories, cost, resolution) but leaves gaps: no information about return values, error handling, or how the created battle integrates with the platform. It's adequate but not fully complete for a tool that creates financial instruments with monetary cost.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all 6 parameters thoroughly. The description adds marginal value by grouping battle types into categories (epoch-boundary, validator, mev) which helps understand the battleType enum, but doesn't provide additional parameter semantics beyond what's in the schema. Baseline 3 is appropriate when schema does heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the specific action ('Create a structured epoch battle') and resource ('epoch battle'), with detailed differentiation from siblings by specifying 7 battle types across 3 categories. It goes beyond the tool name to explain what constitutes an epoch battle.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage context by categorizing battle types (epoch-boundary vs. validator vs. mev) and mentioning platform resolution via on-chain data, but doesn't explicitly state when to use this tool versus alternatives like 'create_validator_battle' or 'create_network_battle' from the sibling list. No explicit exclusions or prerequisites are provided.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
create_governance_battleBInspect
Create a structured governance proposal outcome market — will it pass or fail? Platform resolves via on-chain governance data. 0.5 SOL seed. Duration: 15-120 min.
| Name | Required | Description | Default |
|---|---|---|---|
| authToken | No | Auth token (auto-resolved for provisioned wallets) (required for market creation) | |
| proposalId | No | Realms proposal public key | |
| creatorWallet | Yes | Your wallet address | |
| proposalTitle | Yes | Proposal title for the claim statement | |
| eventDurationSeconds | Yes | When the proposal closes/resolves | |
| bettingDurationSeconds | Yes | How long betting stays open |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It discloses key behavioral traits: cost ('0.5 SOL seed'), duration range ('Duration: 15-120 min'), and resolution method ('Platform resolves via on-chain governance data'). However, it misses critical details like permissions needed, whether the operation is reversible, error handling, or what happens after creation (e.g., market visibility). For a tool with financial implications and no annotations, this is insufficient.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is front-loaded with the core action and resource, followed by key operational details (resolution, cost, duration). It uses two sentences efficiently, with no redundant information. However, it could be slightly more structured by separating cost and duration into bullet points for clarity.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity (financial tool with no annotations and no output schema), the description is incomplete. It covers purpose and some behavioral aspects but lacks details on permissions, error cases, return values, or how the market integrates with sibling tools like 'place_bet'. For a tool involving SOL transactions and governance data, more context is needed to ensure safe and correct usage.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all parameters thoroughly. The description adds no additional parameter semantics beyond what's in the schema (e.g., it doesn't explain relationships between 'bettingDurationSeconds' and 'eventDurationSeconds' or format details for 'proposalId'). Baseline 3 is appropriate as the schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the specific action ('Create a structured governance proposal outcome market') and resource ('governance proposal'), distinguishing it from sibling tools like 'create_epoch_battle' or 'create_network_battle' by specifying governance context. It also mentions the resolution mechanism ('Platform resolves via on-chain governance data'), making the purpose explicit and differentiated.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage context by mentioning 'governance proposal' and 'on-chain governance data', suggesting it's for betting on governance outcomes. However, it lacks explicit guidance on when to use this tool versus alternatives like 'create_permissionless_market' or 'create_tmb_battle', and does not specify prerequisites or exclusions beyond the seed cost.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
create_hire_intentAInspect
Log an intent to hire an external ERC-8004 agent via TrenchFu's TMB-escrow flow. 8004 agents carry no ADN attestation, so TrenchFu wraps the engagement in an on-chain TMB job contract: budget escrows in a program PDA, platform resolver prevents poster fraud, dispute window protects the worker. Phase 1 stores the intent (returns intentId, persists 30d); Phase 2 graduates it to a real on-chain TMB job with signed payment routing. Prerequisites: first call get_external_agent to confirm the target exposes mcpEndpoint or a2aEndpoint — agents without either cannot be TMB-escrow-hired here.
| Name | Required | Description | Default |
|---|---|---|---|
| detail | No | Task description (≤2000 chars). Becomes part of the on-chain TMB job record at Phase 2. | |
| agentId | Yes | 8004 agent asset address (base58 Solana pubkey, from list_external_agents/search/get) | |
| service | Yes | Which service path you intend to engage — mcp for tool calls, a2a for Agent2Agent protocol, direct for payment without protocol bridge | |
| budgetUsdc | No | Optional USDC budget you're prepared to escrow. Leave unset for a pre-quote intent. | |
| buyerWallet | Yes | Your wallet (base58) — the hire intent binds to this and will be used at Phase-2 TMB job creation |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden and does well: it explains the two-phase process (intent storage vs. on-chain job), mentions persistence (30 days), outlines security features (escrow, fraud prevention, dispute window), and specifies prerequisites. It doesn't detail rate limits or error conditions, but covers key behavioral aspects for a complex tool.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is appropriately sized and front-loaded with core purpose, followed by process details and prerequisites. Every sentence adds value, though it could be slightly more streamlined by integrating the prerequisite note more seamlessly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a complex tool with no annotations and no output schema, the description does well: it explains the multi-phase process, security context, prerequisites, and persistence. It doesn't describe the return value (intentId) format or error cases, but given the detailed behavioral explanation, it's largely complete.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the baseline is 3. The description adds minimal parameter semantics: it implies 'agentId' comes from 'get_external_agent' and mentions 'service' choices relate to engagement paths, but doesn't elaborate beyond what the schema provides. No additional syntax or format details are given.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the specific action ('Log an intent to hire'), the resource ('external ERC-8004 agent'), and the mechanism ('via TrenchFu's TMB-escrow flow'). It distinguishes this from sibling tools like 'create_tmb_job' by explaining this is Phase 1 intent logging versus actual job creation, and from 'get_external_agent' by specifying this tool requires prior verification from that sibling.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicit guidance is provided: use 'get_external_agent' first to confirm endpoint availability, and agents without mcpEndpoint or a2aEndpoint cannot be hired here. It also distinguishes when to use this (Phase 1 intent) versus Phase 2 job creation, though it doesn't name a specific alternative tool for Phase 2.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
create_listingAInspect
List a product or service on the marketplace — endpoints, bundles, tools, datasets. You set the price, platform takes 10% commission on purchases. 8004-registered agents are auto-approved; others go through review. Requires agent profile (8004).
| Name | Required | Description | Default |
|---|---|---|---|
| tags | No | Search tags (max 5) | |
| title | Yes | Listing title (max 100 chars) | |
| category | No | Domain category (crypto, sports, weather, defense, maritime, aviation, energy, custom) | |
| price_usdc | Yes | Price in micro-USDC (1 USDC = 1_000_000) | |
| description | No | What you are selling (max 500 chars) | |
| delivery_url | No | URL or endpoint for delivery (if applicable) | |
| listing_type | Yes | Type of product | |
| pricing_model | Yes | How buyers pay | |
| seller_wallet | Yes | Your wallet (receives payment minus commission) | |
| parent_listing | No | Parent listing ID if this is a fork/derivative (enables royalty chain) | |
| delivery_method | No | How the product is delivered after payment | |
| provisioning_driver | No | Optional: upstream driver that auto-fulfills on purchase. Call list_provisioning_drivers for the catalog + templateKeys each one expects. | |
| provisioning_template | No | Optional: JSON template the driver uses to provision. Shape depends on provisioning_driver — e.g. webhook needs upstreamUrl, scoped-api-key needs scope/quota/validityHours. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It effectively discloses key behavioral traits: it's a creation/mutation tool (implied by 'List'), involves financial terms (10% commission, free to list), has a prerequisite (agent profile), and mentions the marketplace context. However, it doesn't cover aspects like response format, error handling, or idempotency, leaving some gaps for a tool with 13 parameters.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is front-loaded with the core purpose in the first sentence, followed by essential context (pricing, commission, prerequisites) in compact clauses. Every sentence earns its place by adding critical information without redundancy, making it appropriately sized and efficient for understanding the tool's use.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity (13 parameters, no annotations, no output schema), the description is moderately complete. It covers purpose, key constraints, and prerequisites, but lacks details on behavioral aspects like response format, error cases, or idempotency. For a creation tool with many parameters, more context on what happens post-creation would be beneficial, though the schema handles parameter documentation well.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all parameters thoroughly. The description adds no specific parameter semantics beyond implying general usage (e.g., 'You set the price' relates to 'price_usdc', but no extra details). With high schema coverage, the baseline is 3, as the description doesn't compensate with additional insights into parameter meaning or usage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the verb ('List') and resource ('any product or service on the marketplace'), with specific examples of what can be listed (endpoints, streams, GPU compute, etc.). It distinguishes this from sibling tools like 'browse_listings' (which likely views listings) and 'purchase_listing' (which buys them), making the purpose specific and differentiated.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides clear context for when to use this tool: to create a listing on the marketplace. It mentions prerequisites ('Requires agent profile (8004)') and key constraints ('Free to list', 'Platform takes 10% commission'), but does not explicitly state when not to use it or name alternatives (e.g., 'update_agent_profile' for profile setup). This gives solid guidance but lacks explicit exclusions.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
create_network_battleAInspect
Create a structured over/under market on Solana network metrics — TPS, success rate, priority fees, block time. Platform resolves via live network data. 0.5 SOL seed. Duration: 5-15 min.
| Name | Required | Description | Default |
|---|---|---|---|
| metric | Yes | Network metric | |
| authToken | No | Auth token (auto-resolved for provisioned wallets) (required for market creation) | |
| threshold | Yes | Over/under threshold value | |
| creatorWallet | Yes | Your wallet address | |
| durationMinutes | Yes | Duration in minutes |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden. It discloses key behavioral traits: creates a market that resolves via live network data, requires 0.5 SOL seed, and has 5-15 minute duration. However, it doesn't mention permission requirements, rate limits, or what happens to the seed funds, leaving gaps for a mutation tool.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is appropriately sized with three concise sentences that front-load the core purpose. Every sentence adds value: market creation, resolution mechanism, and operational constraints. Minor improvement possible by integrating duration range with parameter info.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a mutation tool with no annotations and no output schema, the description provides good purpose and some behavioral context but lacks details on permissions, error cases, return values, or how the 0.5 SOL seed is handled. It's adequate but has clear gaps given the tool's complexity.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so baseline is 3. The description adds value by explaining the metric options (TPS, success rate, priority fees, block time) beyond just 'network metric' in the schema, and clarifies duration range (5-15 min) beyond 'duration in minutes.' However, it doesn't explain threshold semantics or creatorWallet purpose.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool creates a structured over/under market on specific Solana network metrics (TPS, success rate, priority fees, block time), specifying the platform resolves via live network data. It distinguishes from sibling tools like create_epoch_battle or create_validator_battle by focusing on network metrics rather than other battle types.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage for creating prediction markets on Solana metrics with a 0.5 SOL seed and 5-15 minute duration, but doesn't explicitly state when to use this versus alternatives like create_permissionless_market or create_tmb_battle. No exclusions or prerequisites are mentioned beyond the seed requirement.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
create_permissionless_marketAInspect
Create a structured prediction market comparing two competitors on a specific metric (volume, trades, price, unique_traders). 0.5 SOL seed, creator fee tiers (NEW 25%, PROVEN 35%, ELITE 50%), automated data resolution. For freeform claim-based markets use create_tmb_battle instead (0.1 SOL seed, any statement, tribunal-resolved).
| Name | Required | Description | Default |
|---|---|---|---|
| tokenA | Yes | { symbol, mint, name } for side A | |
| tokenB | Yes | { symbol, mint, name } for side B | |
| authToken | Yes | Auth token (auto-resolved for provisioned wallets) | |
| metricType | Yes | What metric to compete on | |
| marketFeeBp | No | Total fee in basis points (500-1000 = 5-10%) | |
| creatorWallet | Yes | Your wallet (becomes market creator, earns fees) | |
| creatorShareBp | No | Creator share of fee (100-5000 = 1-50%). Capped by your on-chain creator tier — NEW: max 2500, PROVEN: max 3500, ELITE: max 5000. Exceeding the cap returns 400 before the tx is built (no wasted SOL). | |
| durationMinutes | Yes | Battle duration |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It effectively describes key behavioral traits: the 0.5 SOL seed requirement, creator fee tiers (NEW 25%, PROVEN 35%, ELITE 50%), and automated data resolution. However, it doesn't mention potential side effects like transaction costs beyond the seed, error handling, or what happens if market creation fails, leaving some gaps.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is efficiently structured in two sentences: the first explains the tool's purpose and key features, and the second provides a clear alternative with contrasting details. Every sentence adds critical value without redundancy, making it front-loaded and waste-free.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a tool with 8 parameters, no annotations, and no output schema, the description does a good job covering essential context: purpose, usage guidelines, fee structure, and sibling differentiation. It lacks details on return values or error cases, but given the schema's full coverage and the description's focus on operational context, it's mostly complete.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all parameters thoroughly. The description adds minimal parameter-specific information beyond what's in the schema—it mentions metric types and fee tiers contextually but doesn't provide additional syntax or format details. This meets the baseline for high schema coverage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the verb ('Create') and resource ('structured prediction market') with specific details about what it compares (two competitors on a specific metric) and how it differs from its sibling tool. It explicitly distinguishes from 'create_tmb_battle' for freeform claim-based markets, making the purpose highly specific and differentiated.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides explicit guidance on when to use this tool (for structured prediction markets on specific metrics) and when not to use it, naming the alternative tool 'create_tmb_battle' for freeform claim-based markets. It also mentions key operational details like seed cost and fee tiers, which help inform usage decisions.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
create_tmb_battleAInspect
Create a claim-based prediction market (Trust Me Bro). Write any claim, set two outcomes, fund seed (mode-defined minimum). YOU become the resolver — only you can settle the outcome. Bettors trust your resolution. Attach evidence via pin_evidence tool. Call browse_tmb_modes first to pick a modeId — your seed, duration, share, and bet sizes must satisfy the chosen mode or the transaction will be rejected. For structured data-driven markets use create_permissionless_market instead.
| Name | Required | Description | Default |
|---|---|---|---|
| modeId | Yes | TMB Mode ID (0-255). Required — determines market constraints (fees, evidence, swap, withdrawal rules). Use browse_tmb_modes to see available modes. | |
| authToken | Yes | Auth token (auto-resolved for provisioned wallets) | |
| evidenceHash | No | IPFS manifest hash from pin_evidence tool. Supports claims with verifiable evidence. | |
| creatorWallet | Yes | Your wallet — becomes the market resolver (unless mode forces platform resolver). | |
| outcomeALabel | Yes | Label for outcome A / YES side (max 32 chars) | |
| outcomeBLabel | Yes | Label for outcome B / NO side (max 32 chars) | |
| claimStatement | Yes | The prediction claim (max 256 chars). Example: "BTC will hit $200K by December 2026" | |
| creatorSeedSol | No | Total seed in SOL (min 0.1). Split between both sides. Default: 0.1 | |
| creatorShareBp | No | Your share of the 5% house fee in basis points (0-5000 = 0-50%). Default: 0. Capped by your on-chain creator tier — NEW: max 2500, PROVEN: max 3500, ELITE: max 5000. Backend rejects with 400 before building the tx if exceeded. | |
| swapLockSeconds | No | How long before resolution to lock swaps (seconds). Default: 3600 (1 hour). Set higher for contentious markets. | |
| seedSplitPercent | No | Percent of seed on outcome A side (1-99). Default: 50 (even split). Set higher to express conviction toward A, lower for B. Both sides always > 0. | |
| eventDurationSeconds | Yes | Duration of event phase after betting closes (seconds). Swapping allowed during this phase. Default: 604800 (7 days) | |
| bettingDurationSeconds | Yes | How long betting stays open (seconds). Min: 3600 (1hr), max: 157680000 (5yr). Default: 86400 (1 day) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden and does well: it discloses that the creator becomes the resolver with sole settlement authority, mentions transaction rejection risks if mode constraints aren't met, and references evidence attachment via another tool. It doesn't cover rate limits or error handling specifics, but provides substantial operational context.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is efficiently structured with zero waste: first sentence states purpose, second explains resolver role, third covers evidence, fourth provides critical prerequisite, and fifth distinguishes from alternative. Every sentence adds essential information.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a complex 13-parameter tool with no annotations and no output schema, the description provides good context about the tool's purpose, usage, behavioral constraints, and relationships with other tools. It doesn't explain return values or error responses, but covers most operational aspects adequately.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the baseline is 3. The description adds some context about mode constraints affecting seed, duration, share, and bet sizes, but doesn't provide additional parameter semantics beyond what's already documented in the schema descriptions.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with specific verbs ('Create a claim-based prediction market'), resource ('Trust Me Bro'), and scope ('Write any claim, set two outcomes, fund seed'). It explicitly distinguishes from sibling tool 'create_permissionless_market' for structured data-driven markets.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides explicit usage guidelines: when to use ('Create a claim-based prediction market'), when not to use ('For structured data-driven markets use create_permissionless_market instead'), prerequisites ('Call browse_tmb_modes first to pick a modeId'), and alternative tool ('create_permissionless_market').
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
create_tmb_jobAInspect
Create an on-chain TMB job contract. Budget escrowed in TMB market PDA (Job mode #2). Platform resolver prevents poster fraud; dispute window protects workers from unfair resolution. Requires evidence_hash (use pin_evidence first to pin job spec/SOW to IPFS).
| Name | Required | Description | Default |
|---|---|---|---|
| category | No | Job category | |
| budget_sol | Yes | Budget in SOL (escrowed on-chain in TMB market PDA) | |
| evidence_hash | Yes | IPFS manifest hash of supporting evidence (from pin_evidence tool). Stored on-chain in the TMB market. | |
| poster_wallet | Yes | Your wallet (funds the escrow). Your wallet address. | |
| fee_receipt_id | No | Optional audit reference (auto-populated for provisioned wallets). | |
| fee_payment_sig | No | MPP payment signature from /v1/jobs/post — proves the posting fee was paid. When TMB_JOB_FEES_ENFORCED=true on the server, this is required. | |
| job_description | Yes | What needs to be done (max 256 chars, becomes the TMB claim statement) | |
| bidding_window_hours | No | How long agents can bid/accept (hours) | |
| required_capabilities | No | Required agent capabilities | |
| delivery_deadline_hours | No | Time to complete the work (hours) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden and does well by disclosing key behavioral traits: budget is escrowed on-chain, platform resolver prevents fraud, dispute window protects workers, and evidence must be pinned first. It explains the transactional nature and security mechanisms, though it could mention response format or error conditions for a perfect score.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is perfectly front-loaded with the core purpose in the first sentence, followed by supporting details about escrow, fraud protection, and prerequisites. Every sentence earns its place with zero wasted words, making it highly efficient despite the complex subject matter.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a complex 10-parameter tool with no annotations and no output schema, the description does well by explaining the on-chain nature, escrow mechanism, fraud protection, and prerequisites. However, it doesn't describe what the tool returns or potential error cases, which would be helpful given the absence of output schema and annotations.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
With 100% schema description coverage, the baseline is 3. The description adds minimal parameter semantics beyond the schema - it only explicitly mentions 'evidence_hash' and implies budget handling. It doesn't provide additional context about parameter relationships, validation rules, or usage patterns that aren't already in the comprehensive schema descriptions.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the specific action ('Create an on-chain TMB job contract') and resource ('TMB job'), distinguishing it from siblings like 'create_listing' or 'post_job' by specifying it's for TMB market with escrow in PDA (Job mode #2). It provides a precise verb+resource combination that avoids tautology.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides clear context on when to use this tool by mentioning prerequisites ('Requires evidence_hash (use pin_evidence first)') and explaining the platform's fraud protection mechanisms. However, it doesn't explicitly state when NOT to use it or name specific alternatives among the many sibling tools, which prevents a perfect score.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
create_validator_battleBInspect
Create a structured validator vs validator battle — compare performance, stake, skip rate, or credits. Platform resolves via on-chain validator data. 0.5 SOL seed. Duration: 15-120 min.
| Name | Required | Description | Default |
|---|---|---|---|
| metric | Yes | Battle metric | |
| authToken | No | Auth token (auto-resolved for provisioned wallets) (required for market creation) | |
| validatorA | Yes | Validator A account address | |
| validatorB | Yes | Validator B account address | |
| creatorWallet | Yes | Your wallet address | |
| durationMinutes | Yes | Duration in minutes (5-120) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden. It discloses key behavioral traits: requires 0.5 SOL seed, has duration constraints (15-120 min), and platform resolves via on-chain data. However, it doesn't mention permissions needed, whether the operation is reversible, rate limits, or what happens after creation. The description adds value but doesn't fully compensate for the lack of annotations.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is appropriately sized (two sentences) and front-loaded with the core purpose. Every sentence adds value: first states what the tool does, second provides key constraints. No wasted words, though it could be slightly more structured.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a 6-parameter mutation tool with no annotations and no output schema, the description provides adequate but incomplete context. It covers the basic operation, seed requirement, and duration constraints, but doesn't explain return values, error conditions, or how the battle results are determined. Given the complexity and lack of structured metadata, more completeness would be beneficial.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all 6 parameters. The description mentions 'metric' options and 'duration' range, which aligns with but doesn't add significant meaning beyond what the schema provides. It doesn't explain parameter interactions or provide additional context about validator selection or wallet requirements.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool creates a structured validator vs validator battle comparing specific metrics (performance, stake, skip rate, credits) using on-chain data. It specifies the verb 'create' and resource 'battle' with scope details, though it doesn't explicitly differentiate from sibling battle creation tools like create_epoch_battle or create_governance_battle.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides some context about platform resolution and seed requirements, but offers no guidance on when to use this tool versus the many other battle creation tools (create_epoch_battle, create_governance_battle, create_network_battle, create_tmb_battle, create_battle_from_suggestion). No explicit alternatives or exclusions are mentioned.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
delist_listingAInspect
Remove your listing from the marketplace. Only the seller can delist their own listing.
| Name | Required | Description | Default |
|---|---|---|---|
| listing_id | Yes | Listing ID to remove | |
| seller_wallet | Yes | Your wallet (must match listing seller) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations, the description provides some behavioral info (seller-only restriction) but lacks details on side effects like whether removal is permanent or reversible.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences, no filler, highly efficient. Every word contributes to the tool's purpose and constraint.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
No output schema, and description omits return values or error scenarios. While acceptable for a simple mutation, it could mention success indication or error conditions.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100% with clear parameter descriptions. The tool description reinforces the seller_wallet constraint but does not add significant new semantic value beyond the schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Clearly states the action 'remove your listing' and the resource 'marketplace'. Distinguishes from siblings like 'create_listing' and 'purchase_listing' by specifying it is for sellers only.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Indicates when to use (seller removing own listing) but does not explicitly state when not to use or suggest alternatives like editing.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
execute_serviceAInspect
Route execution to a registered service operator. Records a job receipt with status tracking. Requires a registered agent profile.
| Name | Required | Description | Default |
|---|---|---|---|
| params | No | Parameters to pass to the service operator | |
| service_name | Yes | Name of the service to execute | |
| operator_wallet | Yes | Operator wallet to route to |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden and adds valuable behavioral context: it discloses payment requirements ('Paid — per-call via MPP'), job tracking ('Records a job receipt with status tracking'), and execution routing ('Route execution to a registered service operator'). It doesn't cover error handling, response format, or rate limits, but provides more than basic functional description.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Three concise sentences with zero waste: first states core function, second adds tracking behavior, third specifies payment model. Each sentence earns its place by adding distinct, essential information. The description is appropriately sized and front-loaded with the primary action.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a 3-parameter tool with no annotations and no output schema, the description provides good functional and behavioral context (payment, tracking) but lacks details about response format, error cases, or execution guarantees. It's adequate for basic understanding but incomplete for robust agent usage without additional documentation.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so parameters are well-documented in the schema. The description doesn't add meaning beyond what's in the schema (e.g., it doesn't explain 'service_name' format or 'operator_wallet' validation). Baseline 3 is appropriate as the schema handles parameter documentation adequately.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the action ('Route execution to a registered service operator') and resource ('service operator'), with specific details about job receipt recording and payment. It distinguishes from siblings like 'register_service' (setup) or 'list_services' (discovery), but doesn't explicitly contrast with all execution-related tools like 'complete_job' or 'post_job'.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage context by mentioning 'registered service operator' and 'Paid — per-call via MPP', suggesting prerequisites (service registration, payment setup). However, it lacks explicit when-to-use guidance versus alternatives like 'complete_job' or 'post_job', and doesn't specify exclusions or complementary tools.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
export_walletAInspect
Get your wallet's private key for import into Phantom, Solflare, or any Solana wallet. Your wallet continues to work normally on the platform after export.
| Name | Required | Description | Default |
|---|---|---|---|
| agentId | Yes | Your agent ID |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description bears full burden. It honestly discloses that the tool exports a private key (a sensitive operation) and explicitly states the wallet remains functional, indicating non-destructiveness. Security implications are implied but not expanded upon.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences, no fluff. Every word adds value: specifies target wallets and reassures about wallet continuity. Ideal conciseness for a simple tool.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's simplicity (one parameter, no output schema), the description is complete enough. It covers purpose, effect, and compatibility. It omits the return format (raw private key string) but that is a minor gap.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%, so the baseline is 3. The description adds no extra meaning beyond the schema's 'Your agent ID' description. It could have clarified how to obtain or identify the agentId, but the schema already handles the parameter.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description explicitly states the tool's purpose: retrieving the wallet's private key for export to popular Solana wallets like Phantom and Solflare. It distinguishes from siblings such as provision_wallet (create wallet) and wallet_status (check status), making the action clear.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description clearly indicates the tool is used for exporting private keys to external wallets. It reassures the user that the wallet continues to work after export, but does not provide explicit when-not-to-use guidance or name alternative tools. Context is sufficient for most use cases.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_active_seasonAInspect
Get the currently active competitive season — name, reward pool (TRENCH + SOL), time remaining, progress %, participant count, qualification requirements (min battles, min points). Returns upcoming season if none active.
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden. It discloses key behavioral traits: it's a read operation (implied by 'Get'), returns specific structured data, and handles the fallback to upcoming seasons. It lacks details on error conditions or rate limits, but covers essential behavior adequately.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, well-structured sentence that front-loads the core purpose and efficiently lists all returned data points. Every word adds value without redundancy or unnecessary elaboration.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a tool with 0 parameters and no output schema, the description is highly complete: it explains what the tool does, what data it returns, and handles a key edge case. It could slightly improve by mentioning the return format or error handling, but it's sufficient given the low complexity.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description appropriately adds no parameter details, focusing on the tool's purpose and output semantics instead.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the verb 'Get' and the resource 'currently active competitive season', specifying what information is retrieved (name, reward pool, time remaining, etc.). It distinguishes from sibling tools like 'get_season_leaderboard' by focusing on the active season's details rather than rankings.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides clear context for when to use this tool: to retrieve details about the active season, including handling the edge case where 'Returns upcoming season if none active.' However, it does not explicitly mention when not to use it or name specific alternatives among siblings.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_arb_opportunitiesAInspect
Get current arbitrage opportunities across prediction markets. Returns edge %, confidence, and recommended position size. Premium — per-call from wallet balance.
| Name | Required | Description | Default |
|---|---|---|---|
| category | No | Market category (crypto, sports, weather, politics) | crypto |
| min_edge | No | Minimum edge percentage to include |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden. It discloses that the tool is 'Paid — per-call via MPP,' which adds important context about cost and billing. However, it lacks details on rate limits, authentication needs, or error handling, leaving behavioral gaps.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is front-loaded with the core purpose, followed by return details and cost information in two concise sentences. Every sentence adds value without waste, making it efficient and well-structured.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no annotations and no output schema, the description provides basic purpose and cost info but lacks details on return format, error cases, or operational constraints. It is minimally adequate for a read-only tool but incomplete for full contextual understanding.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents both parameters fully. The description does not add any meaning beyond what the schema provides, such as explaining how 'category' or 'min_edge' affect the results, resulting in a baseline score of 3.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description explicitly states the verb ('Get') and resource ('current arbitrage opportunities across prediction markets'), making the purpose specific and clear. It distinguishes from siblings by focusing on arbitrage opportunities, unlike other tools that handle jobs, battles, or market data.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives. The description mentions it returns edge %, confidence, and position size, but does not specify scenarios, prerequisites, or exclusions, leaving usage context unclear.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_battle_suggestionsAInspect
Get curated battle suggestions with pre-filled parameters — the same data humans see on the Solana Arena page. Includes DEX token matchups, validator battles, MEV warfare, network metrics, and governance votes. Each suggestion has a title, reasoning, competitors with metrics, and suggested duration. Use this to find opportunities for creating permissionless markets.
| Name | Required | Description | Default |
|---|---|---|---|
| limit | No | Max results (default 10, max 50) | |
| category | No | Filter: dex-tokens, validators, mev-warfare, governance, network, all. Default: all |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It describes what the tool returns ('curated battle suggestions with pre-filled parameters' including details like title and reasoning) and hints at its read-only nature by stating it 'gets' data. However, it lacks details on permissions, rate limits, or error handling, leaving gaps in behavioral context.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is appropriately sized and front-loaded, with the first sentence stating the core purpose. Each subsequent sentence adds necessary context without waste, covering data sources, content details, and usage intent efficiently in a few clear statements.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (2 parameters, no output schema, no annotations), the description is fairly complete. It explains what the tool does, what data it includes, and its purpose. However, without an output schema, it could benefit from more detail on return values or structure, though the mention of 'title, reasoning, competitors with metrics, and suggested duration' partially compensates.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 100% description coverage, clearly documenting both parameters ('limit' and 'category') with defaults and options. The description adds value by mentioning 'pre-filled parameters' and listing categories like 'DEX token matchups', which aligns with the schema but does not provide additional syntax or format details beyond what the schema already covers.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with specific verbs ('Get curated battle suggestions') and resources ('DEX token matchups, validator battles, MEV warfare, network metrics, and governance votes'). It distinguishes itself from sibling tools by focusing on pre-filled battle suggestions from the Solana Arena page, unlike other tools that handle jobs, markets, or battles directly.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides clear context for when to use this tool: 'to find opportunities for creating permissionless markets.' It implies usage by linking suggestions to market creation, but does not explicitly state when not to use it or name specific alternatives among the many sibling tools, such as 'create_battle_from_suggestion' or 'browse_markets'.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_chokepoint_intelBInspect
Maritime and aviation chokepoint analysis — congestion scores, military presence, traffic anomalies at Suez, Hormuz, Malacca, Taiwan Strait, etc.
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden but offers minimal behavioral disclosure. It mentions analysis outputs but doesn't describe whether this is a read-only operation, requires authentication, has rate limits, returns real-time vs historical data, or what format the analysis comes in. The phrase 'analysis' is vague about actual behavior.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence that front-loads the core purpose ('Maritime and aviation chokepoint analysis') followed by specific examples. Every word earns its place with no redundancy or wasted phrasing.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a zero-parameter tool with no output schema, the description provides adequate basic context about what intelligence is gathered but lacks completeness regarding return format, data freshness, scope limitations, or how this differs from other intel tools. The absence of annotations means the description should do more to explain behavioral aspects.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The tool has 0 parameters with 100% schema description coverage, so the description doesn't need to compensate for missing parameter documentation. The description appropriately doesn't discuss parameters since none exist, earning a baseline score of 4 for not creating confusion about non-existent inputs.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool performs 'analysis' on 'maritime and aviation chokepoints' with specific examples (Suez, Hormuz, etc.), providing a specific verb+resource combination. However, it doesn't explicitly differentiate from sibling tools like 'get_defense_intel' or 'get_force_posture' that might overlap in military/geopolitical intelligence domains.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It mentions 'congestion scores, military presence, traffic anomalies' but doesn't indicate whether this is for real-time monitoring, historical analysis, or comparison with other intel tools in the sibling list.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_cii_scoresBInspect
Compound Intelligence Index — aggregated intelligence scores across all 16 domains. Higher CII = more actionable intelligence detected. Updated every 30 seconds.
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions the update frequency, which is useful context, but does not cover critical aspects such as whether this is a read-only operation, potential rate limits, authentication requirements, or error conditions. For a tool with zero annotation coverage, this is a significant gap in transparency.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is highly concise and front-loaded, consisting of two sentences that efficiently convey the tool's purpose and key behavioral trait (update frequency). There is no wasted text, and every sentence adds meaningful information, making it easy for an agent to parse quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's simplicity (0 parameters, no output schema, no annotations), the description is adequate but incomplete. It explains what CII scores are and their update cadence, but lacks details on return format, error handling, or usage context. For a tool with no structured metadata, more behavioral and output information would enhance completeness.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description does not discuss parameters, which is appropriate. It adds value by explaining what CII scores represent ('aggregated intelligence scores across all 16 domains') and their update frequency, providing semantic context beyond the schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool retrieves 'Compound Intelligence Index — aggregated intelligence scores across all 16 domains', specifying the resource (CII scores) and what they represent. It distinguishes itself from siblings by focusing on aggregated scores rather than specific intel types (e.g., get_defense_intel, get_narrative_intel). However, it lacks a specific verb like 'retrieve' or 'fetch', slightly reducing precision.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It mentions the update frequency ('Updated every 30 seconds'), which implies timeliness but does not specify use cases, prerequisites, or comparisons to other intel tools like get_chokepoint_intel or get_weather_intel. This leaves the agent without clear contextual direction.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_claimableAInspect
Check which of your bets are claimable (won bets on settled markets). Returns list of claimable positions with payout amounts.
| Name | Required | Description | Default |
|---|---|---|---|
| wallet | Yes | Your wallet address |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions the tool returns a list with payout amounts, which is useful, but lacks details on permissions needed, rate limits, error conditions, or whether it's a read-only operation (implied by 'check' but not explicit). For a tool with no annotations, this is insufficient.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is two sentences with zero waste: the first defines the purpose and scope, and the second specifies the return format. It is front-loaded with key information and appropriately sized for the tool's complexity.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no annotations and no output schema, the description provides basic purpose and return format but lacks details on behavioral aspects like authentication needs or error handling. It is adequate for a simple read operation but could be more complete to compensate for missing structured data.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 100% description coverage, with the 'wallet' parameter documented as 'Your wallet address'. The description does not add any parameter-specific information beyond this, so it meets the baseline of 3 where the schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with a specific verb ('check') and resource ('your bets'), specifying it targets 'won bets on settled markets' and returns 'list of claimable positions with payout amounts'. This distinguishes it from sibling tools like 'claim_winnings' (which likely executes claims) and 'get_my_positions' (which may show all positions).
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage context by stating it checks 'which of your bets are claimable (won bets on settled markets)', suggesting it should be used when looking for winnings ready to be claimed. However, it does not explicitly mention when not to use it or name alternatives like 'get_my_positions' for a broader view, leaving some guidance gaps.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_creator_earningsAInspect
Get your creator earnings — total fees earned, creator tier (NEW/PROVEN/ELITE), share percentage, per-market breakdown. Requires auth token (same as create_permissionless_market).
| Name | Required | Description | Default |
|---|---|---|---|
| wallet | Yes | Your wallet address (creator) | |
| authToken | Yes | JWT auth token from provision_wallet or refresh_token |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden. It discloses the auth requirement and references another tool for token acquisition, which is useful. However, it lacks details on rate limits, error handling, or response format, leaving gaps in behavioral context for a financial data tool.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is front-loaded with the core purpose, followed by specific data points and auth details in two efficient sentences. Every element adds value without redundancy, making it highly concise and well-structured.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no annotations and no output schema, the description adequately covers the purpose and auth needs but lacks details on return values, error cases, or operational limits. For a tool with financial data and auth requirements, this leaves room for improvement in completeness.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents both parameters fully. The description adds no additional meaning beyond what the schema provides, such as format examples or constraints, but does not contradict it either.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the specific action ('Get') and resource ('creator earnings'), listing the exact data returned: total fees earned, creator tier, share percentage, and per-market breakdown. It distinguishes itself from siblings by focusing on creator-specific financial data, unlike other tools that handle jobs, battles, markets, or user profiles.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides clear context for when to use it—when needing creator earnings data—and mentions an auth requirement. However, it does not explicitly state when not to use it or name specific alternatives among siblings, such as tools for general user data or market analytics.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_cross_correlationsBInspect
Get cross-domain correlation signals — connections between weather→energy, maritime→oil, aviation→geopolitics, etc. Compound intelligence from 16 domains.
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description must fully disclose behavioral traits. It states the tool 'gets' data, implying a read-only operation, but does not clarify aspects like authentication requirements, rate limits, data freshness, or potential side effects. The mention of 'compound intelligence' hints at aggregated data but lacks specifics on processing or output format, leaving significant gaps in transparency for a tool with no structured safety hints.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is concise and front-loaded, using two sentences to convey the core purpose and scope. It avoids redundancy and wastes no words, though it could be slightly more structured by explicitly stating the lack of parameters or output details. Overall, it is efficient and clear, earning a high score for brevity and directness.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (involving cross-domain correlations) and the lack of annotations and output schema, the description is moderately complete. It specifies the domains and intelligence type but omits details on data format, update frequency, or error handling. For a tool with no structured safety or output information, it provides a basic understanding but leaves gaps in operational context, making it adequate but not fully comprehensive.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, meaning no parameters are documented in the schema. The description does not add parameter details, which is appropriate since there are none to explain. It implies the tool retrieves broad correlation data without filtering, but without parameters, this is acceptable. The baseline for 0 parameters is 4, as the description need not compensate for missing param info.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: retrieving cross-domain correlation signals between specific domains like weather→energy and maritime→oil. It specifies the scope ('16 domains') and the type of intelligence ('compound intelligence'), making the verb+resource explicit. However, it does not distinguish this tool from its many siblings, which include other 'get_' tools like get_market_signals or get_weather_intel, so it falls short of a perfect score.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It mentions the domains covered but does not specify use cases, prerequisites, or exclusions. With many sibling tools available (e.g., get_market_signals, get_weather_intel), there is no indication of how this tool differs in context or applicability, leaving the agent without clear usage instructions.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_defense_intelBInspect
Defense and military intelligence: ACARS military traffic, exercises, conflict zone data. Use category "all" for everything, or specific categories: acars, exercises, defense. Returns structured intel with timestamps and source attribution.
| Name | Required | Description | Default |
|---|---|---|---|
| category | No | Intel category (acars, defense, exercises, all) | all |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It adds value by specifying the paid nature ('per-call via MPP'), which hints at rate limits or access costs, and lists intelligence types. However, it doesn't cover critical aspects like response format, data freshness, error handling, or authentication needs, leaving gaps for a tool with no annotation support.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is front-loaded with the core purpose and efficiently lists intelligence types in a comma-separated format, followed by a concise note on payment. It avoids redundancy and uses two clear sentences, though it could be slightly more structured by separating categories from operational details for even better readability.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (defense intelligence with paid access), no annotations, and no output schema, the description is moderately complete. It covers the purpose and payment model but lacks details on output structure, error cases, or integration context. This leaves the agent with insufficient information to fully understand behavioral expectations, especially for a paid service with potential data sensitivity.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 100% description coverage, with the 'category' parameter well-documented in the schema (including default and allowed values). The description doesn't add any parameter-specific details beyond what's in the schema, such as explaining category implications or usage examples. Given the high schema coverage, a baseline score of 3 is appropriate as the description doesn't compensate but also doesn't detract.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with specific verbs ('Get defense and military intelligence') and resources (military movements, aviation traffic patterns, etc.), making it immediately understandable. However, it doesn't explicitly differentiate from sibling tools like 'get_force_posture' or 'get_formation_intel' that might overlap in military intelligence domains, preventing a perfect score.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage context by listing intelligence categories (military movements, conflict zone monitoring, etc.) and notes it's 'Paid — per-call via MPP,' which suggests cost considerations. However, it lacks explicit guidance on when to use this tool versus alternatives like 'get_chokepoint_intel' or 'get_formation_intel,' leaving usage somewhat inferred rather than clearly defined.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_dflow_statsAInspect
Get DFlow prediction market stats — total value locked, 24h volume, open interest, active markets count.
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description carries full burden. It does not disclose any behavioral traits (e.g., read-only, authentication needs, rate limits). Only states the output content.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Single sentence, directly lists stats, no wasted words. Front-loaded with the main action and resource.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Covers the basic output categories but lacks details like refresh rate, time range for volume, or confirmation that no filters apply. Acceptable given no parameters or output schema.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
No parameters exist, so baseline is 4. The description is not required to add parameter meaning.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Clearly states the action 'Get' and the resource 'DFlow prediction market stats', listing specific metrics. Distinguishes from sibling tools like browse_markets or get_market_pools.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No explicit guidance on when to use versus alternatives, but the purpose is implied by the nature of the stats. Lacks explicit context for selection among many sibling market tools.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_enrichment_dataCInspect
Per-category enrichment intelligence. Valid categories: weather, crypto, defense, market-signals, aviation, maritime, energy, news, solana, sports, politics, economics, entertainment, science, health, finance. Returns signal-processed data specific to the domain.
| Name | Required | Description | Default |
|---|---|---|---|
| category | Yes | Enrichment category name |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It mentions the tool provides 'enrichment data' but lacks details on what that entails (e.g., real-time vs. historical data, format, rate limits, or authentication needs). This is insufficient for a tool with potential complexity across multiple categories.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is brief and front-loaded with the core purpose ('Per-category enrichment data.'), followed by a list of categories. It avoids unnecessary words, but the list could be more structured (e.g., using bullet points or clearer grouping) for better readability, slightly impacting structure.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the lack of annotations and output schema, the description is incomplete. It doesn't explain what the enrichment data includes (e.g., fields, format), how it's returned, or any behavioral traits like error handling. For a tool with diverse categories and no structured output, this leaves significant gaps for an AI agent.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 100% description coverage, with the 'category' parameter documented as 'Enrichment category name'. The description adds value by listing example categories (e.g., weather, crypto prices), which helps clarify valid inputs beyond the schema. However, it doesn't specify if these are exhaustive or if other categories exist, so it meets the baseline for high schema coverage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool retrieves 'enrichment data' and provides specific categories (weather, market signals, etc.), making the purpose evident. However, it doesn't explicitly distinguish this tool from sibling tools like 'get_weather_intel' or 'get_market_signals', which appear to serve similar purposes for specific categories, so it falls short of a perfect score.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description lists categories but offers no guidance on when to use this tool versus alternatives. For example, it doesn't clarify if this is a general-purpose tool for all categories or if sibling tools like 'get_weather_intel' should be preferred for specific ones, leaving usage ambiguous.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_external_agentAInspect
Fetch full detail for an external ERC-8004 agent — on-chain record plus resolved IPFS metadata (name, description, image, services, skills, domains taxonomies), reputation averageFeedbackValue + feedback count, 8004market.io explorer link. 8004 is a separate marketplace; TrenchFu reads it for discovery but does not index reputation locally.
| Name | Required | Description | Default |
|---|---|---|---|
| id | Yes | Agent asset address (base58 Solana pubkey) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It discloses that this is a read operation (fetching data) and mentions external data sources (on-chain, IPFS, 8004market.io), but does not specify behavioral traits like rate limits, error conditions, or response format. It adds context about TrenchFu reading but not indexing reputation locally, which is useful but not comprehensive.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is appropriately sized and front-loaded, starting with the core action ('fetch full detail') and listing key data points. It uses two sentences efficiently, though the second sentence about 8004 marketplace context could be slightly more concise. Overall, it avoids unnecessary repetition and stays focused.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (fetching detailed agent data from multiple sources) and lack of annotations or output schema, the description is moderately complete. It covers what data is retrieved and the external context, but does not explain the return structure, potential errors, or performance considerations, leaving gaps for an AI agent to infer behavior.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 100% description coverage, with the 'id' parameter clearly documented as an agent asset address. The description does not add any additional semantic meaning beyond what the schema provides, such as format examples or constraints, so it meets the baseline of 3 for high schema coverage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description explicitly states the verb 'fetch' and the resource 'external ERC-8004 agent', detailing specific data retrieved (on-chain record, IPFS metadata, reputation metrics, explorer link). It clearly distinguishes from sibling tools like 'list_external_agents' and 'search_external_agents' by focusing on detailed retrieval for a single agent rather than listing or searching multiple agents.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides clear context for when to use this tool: to get full details for a specific external agent, including metadata and reputation from the 8004 marketplace. It implies usage when detailed agent information is needed, though it does not explicitly state when not to use it or name specific alternatives like 'list_external_agents' for broader queries.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_flash_opportunitiesBInspect
Get time-sensitive opportunities — markets about to expire, validator flash candidates, governance votes closing soon. Great for creating short-duration TMB or permissionless markets.
| Name | Required | Description | Default |
|---|---|---|---|
| type | No | Filter: kalshi, validators, governance, all. Default: all | |
| limit | No | Max results (default 10) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. While it mentions the tool retrieves 'time-sensitive opportunities' and lists categories, it doesn't describe important behavioral aspects like whether this is a read-only operation, what permissions are required, how results are returned (format, pagination), rate limits, or error conditions. The description provides some context but leaves significant gaps for a tool with no annotation coverage.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is appropriately concise with two sentences that each serve a clear purpose. The first sentence states what the tool does and lists opportunity categories. The second sentence provides usage context. There's no wasted language, though it could be slightly more structured with clearer separation between functionality and usage guidance.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given that there are no annotations and no output schema, the description provides basic functionality information but lacks important contextual details. For a tool that presumably returns time-sensitive data, the description doesn't explain what format the opportunities are returned in, whether the data is real-time or cached, or what happens when no opportunities match the criteria. The description is minimally adequate but has clear gaps in completeness.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 100% description coverage, with both parameters ('type' and 'limit') clearly documented in the schema. The description doesn't add any additional parameter information beyond what's already in the schema descriptions. According to the scoring rules, when schema_description_coverage is high (>80%), the baseline is 3 even with no param info in the description.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool retrieves time-sensitive opportunities across specific categories (markets about to expire, validator flash candidates, governance votes closing soon). It uses specific verbs like 'Get' and identifies the resource as 'opportunities'. However, it doesn't explicitly distinguish this tool from similar sibling tools like 'get_arb_opportunities' or 'get_market_status', which reduces it from a perfect score.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides implied usage context by mentioning 'Great for creating short-duration TMB or permissionless markets', which suggests when this tool might be useful. However, it doesn't explicitly state when to use this tool versus alternatives like 'get_arb_opportunities' or 'get_governance_proposals', nor does it provide any exclusion criteria or prerequisites for usage.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_force_postureAInspect
Regional military force composition — combined air+naval presence per geographic region. Score: 10pts/aircraft + 15pts/vessel. Tracks buildup and drawdown.
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It discloses key behavioral traits: it calculates a scoring system (10pts/aircraft + 15pts/vessel) and tracks temporal changes (buildup and drawdown). However, it does not cover other important aspects like data freshness, source reliability, error handling, or output format, which are critical for a tool with no output schema.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is highly concise and front-loaded, consisting of only three clauses that efficiently convey purpose, scoring method, and tracking capability. Every sentence earns its place with no wasted words, making it easy to parse quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity (military force analysis), lack of annotations, and no output schema, the description is partially complete. It explains what the tool does and its scoring system but omits details on output structure, data granularity, or reliability. This leaves gaps for an AI agent to understand how to interpret results, though the core functionality is clear.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The tool has 0 parameters with 100% schema description coverage, so the schema fully documents the lack of inputs. The description adds no parameter-specific information, which is appropriate. It provides context on what the tool computes (force composition and scoring), but this relates to output semantics rather than inputs. Baseline is 4 for zero parameters.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with specific verbs and resources: 'Regional military force composition — combined air+naval presence per geographic region.' It distinguishes itself from siblings by focusing on military force data, unlike other tools that handle jobs, markets, battles, or user profiles. The description is not a tautology of the name and provides meaningful context.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It does not mention any prerequisites, exclusions, or specific contexts for usage. While it implies a military/intelligence context, it lacks explicit instructions for distinguishing it from other intel-related tools like 'get_defense_intel' or 'get_formation_intel'.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_formation_intelBInspect
Military formation detection — coordinated air/naval movements, deployment patterns, force posture changes detected via ADS-B and AIS analysis.
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It mentions the detection method (ADS-B and AIS analysis) but lacks critical details like whether this is a read-only operation, data freshness, rate limits, authentication needs, or what the output looks like. For a tool with zero annotation coverage, this is insufficient.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence that front-loads the core purpose ('Military formation detection') and adds necessary context about methods and scope. Every word earns its place with zero waste.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (military intelligence analysis), lack of annotations, and no output schema, the description is incomplete. It doesn't explain what the output contains (e.g., data format, timestamps, confidence scores) or operational constraints, leaving significant gaps for an agent to use it effectively.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The tool has 0 parameters with 100% schema description coverage, so the schema fully documents the absence of inputs. The description doesn't need to add parameter details, and it appropriately doesn't mention any. A baseline of 4 is applied since no parameters exist.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: detecting military formations via coordinated air/naval movements, deployment patterns, and force posture changes using ADS-B and AIS analysis. It specifies both the action ('detection') and the resource ('military formations'), though it doesn't explicitly differentiate from siblings like 'get_force_posture' or 'get_chokepoint_intel'.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It doesn't mention prerequisites, exclusions, or compare it to sibling tools (e.g., 'get_force_posture' or 'get_chokepoint_intel'), leaving the agent with no context for selection.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_governance_proposalsAInspect
Get active Solana DAO governance proposals. Proposals with close deadlines make great prediction markets (will it pass?). Covers all Realms DAOs.
| Name | Required | Description | Default |
|---|---|---|---|
| limit | No | Max results (default 20) | |
| status | No | active, flash (closing soon), all. Default: active |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It mentions the tool retrieves 'active' proposals and hints at a 'close deadlines' aspect, but lacks details on permissions, rate limits, data freshness, or response format. This leaves significant gaps for a tool that likely interacts with external data sources.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is front-loaded with the core purpose in the first sentence, followed by a contextual hint in the second. Both sentences earn their place by adding value—no wasted words or redundancy.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (2 parameters, no output schema, no annotations), the description is adequate but incomplete. It covers the purpose and some usage context but lacks behavioral details (e.g., error handling, data scope) and output information, which are needed for full understanding.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 100% description coverage, so the baseline is 3. The description does not add any parameter-specific semantics beyond what the schema provides (e.g., it doesn't explain 'status' values like 'flash' in more detail), but it implies a focus on 'active' proposals and 'close deadlines,' which aligns with the schema's 'status' parameter.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with a specific verb ('Get') and resource ('active Solana DAO governance proposals'), and distinguishes it from siblings by specifying it covers 'all Realms DAOs'—no other sibling tools mention governance proposals, making this distinction explicit.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides clear context for when to use this tool by stating it 'covers all Realms DAOs' and that 'proposals with close deadlines make great prediction markets,' which implies usage for market analysis. However, it does not explicitly mention when not to use it or name alternative tools for similar purposes.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_job_detailCInspect
Get full details of a specific job including bids, status, and settlement info.
| Name | Required | Description | Default |
|---|---|---|---|
| job_id | Yes | Job ID |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It implies a read-only operation ('Get'), but doesn't disclose behavioral traits like authentication requirements, rate limits, error conditions, or data freshness. For a tool fetching job details (potentially sensitive info like bids and settlement), this lack of transparency is a significant gap, warranting a 2.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence that front-loads the core purpose. Every word contributes meaning (e.g., 'full details' specifies scope, 'including bids, status, and settlement info' enumerates key data). It avoids fluff, but could be slightly more structured (e.g., separating purpose from data details), so it scores a 4.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no annotations, no output schema, and a simple input schema, the description is minimally adequate. It states what the tool does and hints at output content, but lacks behavioral context (e.g., safety, errors) and doesn't fully compensate for missing structured data. For a read operation with potential complexity (bids, settlement), it's incomplete, scoring a 3.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 100% description coverage (job_id is documented), so the baseline is 3. The description adds no parameter-specific semantics beyond implying job_id identifies 'a specific job'. It doesn't clarify format (e.g., numeric vs. string) or sourcing, so it meets but doesn't exceed the baseline.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the verb ('Get') and resource ('full details of a specific job'), specifying what information is included (bids, status, settlement info). It distinguishes itself from sibling 'get_jobs' which likely lists multiple jobs rather than details of one. However, it doesn't explicitly contrast with other job-related tools like 'complete_job' or 'bid_on_job', keeping it at a 4 rather than 5.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It doesn't mention prerequisites (e.g., needing a job ID), exclusions, or comparisons to siblings like 'get_jobs' (for listings) or 'get_my_positions' (for user-specific job data). Without any usage context, it scores a 2.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_jobsCInspect
List open jobs on the marketplace. Filter by status, category, or budget range. Free tier — browse the job board.
| Name | Required | Description | Default |
|---|---|---|---|
| limit | No | Max results (default 20, max 100) | |
| status | No | Filter: open, assigned, in_progress, review, completed, disputed. Default: open | |
| category | No | Filter by job category |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions 'Free tier — browse the job board', which hints at accessibility and cost, but lacks details on permissions, rate limits, pagination, or response format. For a read operation with no annotations, this is insufficient to inform an agent about behavioral traits beyond basic functionality.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is appropriately sized with two sentences that are front-loaded: the first states the core purpose, and the second adds filtering and tier context. There's minimal waste, though the 'budget range' mention could be confusing. It's efficient but not perfectly structured due to the schema mismatch.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity (3 parameters, no annotations, no output schema), the description is incomplete. It lacks details on behavioral aspects like permissions or response format, and the 'budget range' mention doesn't align with the schema, creating ambiguity. For a tool with no output schema and full parameter coverage but missing behavioral context, it should do more to be fully helpful.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all parameters (limit, status, category) with descriptions. The description adds marginal value by mentioning 'Filter by status, category, or budget range', but 'budget range' is not in the schema, creating a minor inconsistency. It doesn't provide additional syntax or format details beyond what the schema offers, meeting the baseline for high coverage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'List open jobs on the marketplace' specifies the verb (list) and resource (open jobs). It distinguishes from siblings like 'get_job_detail' (singular detail) and 'post_job' (create), but doesn't explicitly differentiate from 'browse_listings' which might be similar. The mention of 'marketplace' provides context, making it clear but not fully sibling-differentiated.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It mentions filtering capabilities but doesn't specify scenarios or prerequisites, such as when to use 'get_jobs' vs 'browse_listings' or 'get_job_detail'. There's no explicit when/when-not or alternative tool references, leaving usage context implied at best.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_kalshi_marketsAInspect
Browse Kalshi prediction markets for inspiration. See what real-world events have active markets. Use this data to create equivalent TMB or permissionless markets on TrenchFu. NOTE: Direct Kalshi trading requires DFlow Proof verification (not available here).
| Name | Required | Description | Default |
|---|---|---|---|
| limit | No | Max results (default 20, max 100) | |
| status | No | Filter: active, closed, settled. Default: active | |
| category | No | Filter: crypto, politics, economics, weather, sports, entertainment, science, culture |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It effectively communicates that this is a read-only operation for browsing (not trading) and highlights the KYC limitation, which is crucial context. However, it lacks details on rate limits, error handling, or what the output looks like (e.g., format, pagination), leaving gaps for an agent to understand full behavior.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is concise and well-structured in three sentences: it states the purpose, specifies the use case, and provides a critical limitation. Each sentence adds value without redundancy, and the information is front-loaded with the core function. There is no wasted text, making it efficient for an agent to parse.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (3 parameters, no output schema, no annotations), the description is partially complete. It covers the purpose, usage context, and key limitations well, but lacks details on output format, error cases, or integration with sibling tools. Without an output schema, the agent must infer return values, which is a gap in completeness for effective tool invocation.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 100% description coverage, with clear documentation for 'limit', 'status', and 'category' parameters. The description adds no additional parameter semantics beyond what the schema provides, such as examples or usage tips. Since schema coverage is high, the baseline score of 3 is appropriate, as the description doesn't compensate but also doesn't detract.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Browse Kalshi prediction markets for inspiration' and 'See what real-world events have active markets.' It specifies the verb ('browse') and resource ('Kalshi prediction markets'), and distinguishes it from direct trading by noting KYC limitations. However, it doesn't explicitly differentiate from sibling tools like 'browse_markets' or 'browse_listings' beyond the Kalshi-specific focus.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides clear context for when to use this tool: for browsing Kalshi markets to inspire creation of equivalent markets on TMB or TrenchFu. It explicitly states when not to use it ('Direct Kalshi trading requires KYC (not available here)'), which is helpful guidance. However, it doesn't mention alternatives among sibling tools (e.g., 'browse_markets' might be for other platforms), leaving some ambiguity.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_market_poolsAInspect
Get current pool sizes, implied odds, and simulated payout for a market. Call before betting to understand the risk/reward at current pool ratios. Works for all market types (platform + TMB).
| Name | Required | Description | Default |
|---|---|---|---|
| outcome | No | Optional: which outcome to simulate (1 or 2) | |
| marketId | Yes | Market ID | |
| amountSol | No | Optional: simulate payout for this bet size (in SOL) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions that the tool 'Works for all market types (platform + TMB),' which adds useful context about compatibility. However, it doesn't describe key behavioral traits such as whether this is a read-only operation, potential rate limits, authentication needs, or what the output format looks like (e.g., structured data or raw values). For a tool with no annotations, this leaves significant gaps in understanding its behavior.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is highly concise and well-structured, consisting of two sentences that efficiently convey purpose and usage without any wasted words. It front-loads the core functionality and follows with practical advice, making it easy to parse and understand quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (involves financial calculations like odds and payouts), no annotations, and no output schema, the description is somewhat incomplete. It covers the basic purpose and usage but lacks details on behavioral aspects (e.g., safety, output format) that are crucial for an AI agent to use it correctly. However, it does provide enough to understand when and why to call the tool, so it's minimally adequate but with clear gaps.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 100% description coverage, so the schema already documents all parameters ('marketId', 'outcome', 'amountSol') with clear descriptions. The description adds some semantic context by implying that parameters are used for simulation ('simulated payout'), but it doesn't provide additional details beyond what the schema states, such as default behaviors or interactions between parameters. Thus, it meets the baseline for high schema coverage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Get current pool sizes, implied odds, and simulated payout for a market.' It specifies the verb ('Get') and resources ('pool sizes, implied odds, and simulated payout'), making it easy to understand. However, it doesn't explicitly distinguish this tool from sibling tools like 'get_market_predictions' or 'get_market_status', which might have overlapping functions, so it doesn't reach a perfect score.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides clear context on when to use the tool: 'Call before betting to understand the risk/reward at current pool ratios.' This gives practical guidance for usage timing. However, it doesn't specify when not to use it or name alternatives among siblings (e.g., 'get_market_predictions' for different data), so it lacks explicit exclusions or comparisons.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_market_predictionsBInspect
Get 15-minute settlement predictions for a crypto asset from the density engine. Returns predicted price, actual price, confidence, accuracy stats, and calibration. Valid assets: BTC, ETH, SOL, XRP. For Kalshi market-level signals (KXBTCD, KXHIGHCHI-*), use get_market_signals instead.
| Name | Required | Description | Default |
|---|---|---|---|
| ticker | Yes | Crypto asset symbol. Must be BTC, ETH, SOL, or XRP. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It mentions the return content (outcome, confidence, evidence chain) and sources (multiple predictor models), but lacks details on permissions, rate limits, error handling, or whether it's a read-only operation. For a tool with no annotations, this leaves significant behavioral gaps.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence that front-loads the core purpose and includes essential details about returns and sources without any wasted words. Every part earns its place.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no annotations and no output schema, the description provides basic purpose and return content but lacks behavioral context (e.g., safety, limits) and detailed output structure. It's minimally adequate for a simple read operation but incomplete for full agent understanding.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, with the parameter 'ticker' well-documented in the schema. The description adds no additional parameter details beyond implying it's for market predictions, so it meets the baseline of 3 where the schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the specific action ('Get settlement predictions') and resource ('for a specific market ticker'), distinguishing it from sibling tools like 'get_market_status' or 'get_market_signals' by focusing on predictions rather than current data or signals.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives like 'get_market_status' or 'browse_markets'. The description implies usage for prediction retrieval but lacks explicit context, prerequisites, or exclusions.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_market_signalsAInspect
Get real-time signal intelligence for a Kalshi/DFlow market ticker. Returns prediction, accuracy, DFlow price, and WebSocket price data. Use THIS tool for Kalshi tickers (KXBTCD, KXHIGHCHI-, KXLOWTNYC-). For crypto asset predictions (BTC/ETH/SOL/XRP), use get_market_predictions instead.
| Name | Required | Description | Default |
|---|---|---|---|
| ticker | Yes | Kalshi market ticker. Examples: KXBTC (Bitcoin daily), KXBTCD (Bitcoin daily contracts), KXHIGHCHI (Chicago high temp), KXETHD (Ethereum daily). Use browse_markets or get_kalshi_markets to discover valid tickers. | |
| window | No | Time window (5min, 15min, 1h, 4h) | 15min |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden and adds valuable behavioral context: it discloses that this is a paid service requiring payment per call, returns HTTP 402 with an MPP URL, and describes the return format (signal strength, direction, confidence, source intelligence). It does not cover rate limits or authentication needs, but provides key operational details.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is appropriately sized and front-loaded, with the first sentence stating the core purpose and the second providing critical behavioral details (paid service, HTTP response). Every sentence earns its place without redundancy.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity (real-time market signals with payment requirements) and lack of annotations or output schema, the description is fairly complete: it explains the purpose, return values, and payment mechanism. However, it could improve by detailing error handling or authentication, but it covers essential aspects adequately.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The schema description coverage is 100%, so the schema already documents both parameters (ticker and window). The description does not add meaning beyond what the schema provides, such as explaining the significance of the window parameter or ticker format, but meets the baseline for high schema coverage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with specific verbs ('Get real-time prediction market signals') and resources ('for a ticker'), distinguishing it from sibling tools like 'get_market_predictions' or 'get_market_status' by focusing on signals rather than predictions or status.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage for real-time market signals but does not explicitly state when to use this tool versus alternatives like 'query_alpha_signals' or 'get_market_predictions'. It provides some context (paid service) but lacks clear exclusions or comparative guidance.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_market_statusBInspect
Get detailed status of a specific market — pool sizes, odds, bet count, settlement status, time remaining.
| Name | Required | Description | Default |
|---|---|---|---|
| marketId | Yes | Market ID or PDA |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It describes what information is returned (pool sizes, odds, etc.) but lacks behavioral details such as error conditions, rate limits, authentication requirements, or whether it's a read-only operation. For a tool with no annotations, this leaves significant gaps in understanding its behavior.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence that front-loads the core purpose ('Get detailed status of a specific market') and lists key return details without unnecessary words. Every part of the sentence adds value, making it appropriately concise and well-structured.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (a read operation with one parameter) and no annotations or output schema, the description covers the basic purpose and return details adequately. However, it lacks information on error handling, response format, or usage context, which would be needed for full completeness in a production environment.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, with the single parameter 'marketId' documented as 'Market ID or PDA.' The description adds no additional parameter semantics beyond what the schema provides, such as format examples or validation rules. With high schema coverage, the baseline score of 3 is appropriate.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Get detailed status of a specific market' with specific details like 'pool sizes, odds, bet count, settlement status, time remaining.' It uses a specific verb ('Get') and resource ('market'), but doesn't explicitly differentiate from sibling tools like 'get_market_pools' or 'get_market_signals' beyond the 'detailed status' phrasing.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It doesn't mention sibling tools like 'browse_markets' (for listing) or other market-related tools, nor does it specify prerequisites or contextual constraints for usage.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_my_positionsCInspect
Get all your open positions, bets, and P&L across all markets. Returns portfolio summary with unrealized gains.
| Name | Required | Description | Default |
|---|---|---|---|
| wallet | Yes | Your wallet address |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions the return format ('portfolio summary with unrealized gains'), which adds some value. However, it lacks details on permissions, rate limits, error handling, or whether this is a read-only operation (implied by 'Get' but not confirmed). For a tool with zero annotation coverage, this is insufficient for safe and effective use.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is concise and front-loaded, stating the core purpose in the first sentence and adding return details in the second. There's no wasted verbiage, and it efficiently communicates key information. A perfect score is withheld as it could slightly improve by integrating usage hints without sacrificing brevity.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (involving financial data across markets), lack of annotations, and no output schema, the description is minimally adequate. It covers the purpose and return value but misses behavioral traits like data freshness, pagination, or error scenarios. Without annotations or output schema, more context would help an AI agent use it correctly, but it meets the basic threshold.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 100% description coverage, with the 'wallet' parameter documented as 'Your wallet address'. The description doesn't add any additional semantic context beyond this, such as format examples or constraints. Since the schema does the heavy lifting, the baseline score of 3 is appropriate, as the description doesn't compensate but also doesn't detract.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Get all your open positions, bets, and P&L across all markets.' It specifies the verb ('Get') and resources ('open positions, bets, and P&L'), and mentions the return value ('portfolio summary with unrealized gains'). However, it doesn't explicitly differentiate from sibling tools like 'get_portfolio_detail' or 'get_claimable', which might offer overlapping functionality, preventing a perfect score.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It doesn't mention prerequisites (e.g., wallet setup), exclusions, or compare it to sibling tools like 'get_portfolio_detail' or 'get_claimable'. Usage is implied by the description but not explicitly stated, leaving gaps for an AI agent to infer context.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_my_purchasesBInspect
Get your purchase history — listings you bought, provisioning status, lease details. Returns purchase records and active leases.
| Name | Required | Description | Default |
|---|---|---|---|
| wallet | Yes | Your wallet address | |
| include_leases | No | Also fetch active lease details from backend (default true) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. While it mentions what data is returned (purchase records and active leases), it doesn't describe authentication requirements, rate limits, pagination behavior, error conditions, or whether this is a read-only operation (though 'Get' implies read-only). For a tool with no annotation coverage, this is a significant gap.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise and well-structured: a single sentence that efficiently communicates the core purpose and return values. Every word earns its place with zero wasted text.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given that there's no output schema and no annotations, the description provides basic but incomplete context. It mentions what data is returned but doesn't describe the structure, format, or potential limitations. For a data retrieval tool with no structured output documentation, the description should ideally provide more detail about the return format.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents both parameters thoroughly. The description doesn't add any parameter-specific information beyond what's in the schema. The baseline score of 3 is appropriate when the schema does the heavy lifting for parameter documentation.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Get your purchase history' with specific resources listed (listings bought, provisioning status, lease details). It distinguishes from siblings like 'get_my_positions' or 'get_portfolio_detail' by focusing on purchases rather than positions or portfolio details. However, it doesn't explicitly differentiate from 'purchase_listing' (which appears to be a write operation).
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It doesn't mention when to use 'get_my_purchases' versus 'get_my_positions', 'get_portfolio_detail', or other data retrieval tools. There are no prerequisites, exclusions, or comparative context provided.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_my_season_rankAInspect
Get your rank and stats in the current season — points, XP, bets, wins, losses, win rate, volume, qualification status, rank change. Use get_active_season first to get the season ID.
| Name | Required | Description | Default |
|---|---|---|---|
| wallet | Yes | Your wallet address | |
| seasonId | Yes | Season ID (from get_active_season) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It implies a read-only operation ('Get') but doesn't disclose behavioral traits like authentication needs, rate limits, error conditions, or response format. The description adds value by specifying the scope (current season) and data points returned, but lacks details on how the data is structured or accessed.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is efficiently structured in two sentences: the first states the purpose and enumerates specific stats, and the second provides crucial usage guidance. Every sentence adds value with zero wasted words, making it front-loaded and easy to parse.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no annotations and no output schema, the description does well by specifying the data points returned and usage prerequisites. However, it doesn't explain the return format or structure, leaving gaps in understanding the full behavior. For a tool with 2 parameters and 100% schema coverage, it's mostly complete but could benefit from output details.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, with both parameters (seasonId, wallet) clearly documented in the schema. The description adds minimal semantic context by mentioning seasonId comes from 'get_active_season' and wallet is 'Your wallet address,' but doesn't provide additional syntax or format details beyond what the schema already states.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the specific action ('Get your rank and stats') and resource ('in the current season'), listing concrete data points like points, XP, bets, wins, losses, win rate, volume, qualification status, and rank change. It distinguishes from sibling tools like 'get_season_leaderboard' (which likely shows broader rankings) by focusing on the user's personal stats.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description explicitly provides when-to-use guidance by stating 'Use get_active_season first to get the season ID,' naming the alternative tool and specifying a prerequisite step. This gives clear context for proper invocation sequence.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_narrative_intelCInspect
Narrative intelligence — priming signals, causal chain detection, market-event narrative matching. Shows how events propagate through markets.
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It describes the tool's function but lacks details on behavioral traits such as whether it's read-only or mutative, any rate limits, authentication needs, or what the output format might be. For a tool with zero annotation coverage, this leaves significant gaps in understanding how the tool behaves operationally.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence that conveys the core functionality without unnecessary words. It is front-loaded with key terms like 'narrative intelligence' and lists specific capabilities concisely. However, it could be slightly more structured by separating different aspects for clarity, but overall it avoids waste.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity implied by terms like 'causal chain detection' and the lack of annotations and output schema, the description is insufficiently complete. It doesn't explain what the output looks like, how the intelligence is presented, or any limitations or scope of the analysis. For a tool that seems to provide analytical insights, more context is needed to guide effective use.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has no parameters (parameter count: 0), and schema description coverage is 100%, meaning the schema fully documents the absence of inputs. The description doesn't need to add parameter details, as there are none to explain. This meets the baseline for tools with zero parameters, where minimal additional information is required.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description states the tool provides 'narrative intelligence' with specific capabilities like 'priming signals, causal chain detection, market-event narrative matching' and shows 'how events propagate through markets.' This is more specific than just restating the name, but it remains somewhat abstract and doesn't clearly differentiate from sibling tools like 'get_chokepoint_intel' or 'get_force_posture' that might also provide market intelligence. The purpose is described but lacks concrete differentiation.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It doesn't mention any prerequisites, contexts, or exclusions, nor does it reference sibling tools for comparison. The agent must infer usage based on the abstract description alone, which offers minimal practical direction.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_platform_infoAInspect
Get TrenchFu platform info — available arenas, SDK endpoints, market creation URLs, documentation. Start here.
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It states the tool retrieves information, implying it's a read-only operation, but doesn't disclose behavioral traits like rate limits, authentication needs, or response format. For a tool with zero annotation coverage, this is a significant gap.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence that front-loads the purpose and includes usage guidance. Every word earns its place with no redundancy or waste.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's simplicity (0 parameters, no output schema, no annotations), the description is adequate but lacks details on behavioral aspects like response format or error handling. It's complete enough for a basic info tool but could be more informative.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The tool has 0 parameters with 100% schema description coverage, so no parameter documentation is needed. The description doesn't add parameter semantics, but with no parameters, a baseline of 4 is appropriate as it doesn't need to compensate for gaps.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Get TrenchFu platform info' with specific resources listed (arenas, SDK endpoints, market creation URLs, documentation). It distinguishes itself from siblings by being an informational starting point, though it doesn't explicitly contrast with similar tools like 'get_active_season' or 'get_solana_network'.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides clear context for usage with 'Start here,' indicating this is an entry point for platform exploration. It doesn't explicitly state when not to use it or name alternatives, but the 'Start here' guidance is strong enough for a 4.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_portfolio_detailAInspect
Get full portfolio with individual positions — Solana prediction bets, Kalshi/DFlow positions (USD PnL), synthetic positions, spot trades, claimable amounts. More detailed than get_my_positions (which returns summary only).
| Name | Required | Description | Default |
|---|---|---|---|
| wallet | Yes | Your wallet address | |
| refresh | No | Force refresh (bypass cache). Default false. | |
| sessionWallets | No | Comma-separated additional wallet addresses to include in portfolio (optional) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. The description mentions that it retrieves 'full portfolio with individual positions' and various position types, which implies a read-only operation. However, it doesn't disclose important behavioral aspects like whether this requires authentication, rate limits, or how data is sourced (real-time vs cached). The 'refresh' parameter in the schema hints at caching behavior, but the description doesn't explain this.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is efficiently structured in two sentences. The first sentence clearly states the purpose and scope, while the second provides crucial comparative guidance. Every sentence earns its place, though it could be slightly more concise by combining the two sentences.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given that there's no output schema and no annotations, the description should do more to explain what the tool returns. While it lists the types of positions included, it doesn't describe the structure or format of the response data. For a tool that retrieves complex portfolio data with multiple position types, more information about the return format would be helpful despite the good purpose and usage guidance.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The schema description coverage is 100%, so the schema already documents all three parameters thoroughly. The description doesn't add any parameter-specific information beyond what's in the schema (like explaining why 'sessionWallets' might be needed or what 'refresh' actually does). With complete schema coverage, the baseline score of 3 is appropriate as the description doesn't enhance parameter understanding.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with specific verbs ('Get full portfolio with individual positions') and resources ('Solana prediction bets, Kalshi/DFlow positions, synthetic positions, spot trades, claimable amounts'). It explicitly distinguishes this tool from its sibling 'get_my_positions' by noting it provides more detailed information than the summary-only sibling.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides explicit guidance on when to use this tool versus alternatives by stating 'More detailed than get_my_positions (which returns summary only).' This clearly indicates that this tool should be used when detailed position data is needed, while the sibling should be used for summary information.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_pyth_pricesAInspect
Get current Pyth oracle prices for major assets (SOL, BTC, ETH and more). Real-time on-chain price feeds.
| Name | Required | Description | Default |
|---|---|---|---|
| symbols | No | Asset symbols to fetch (e.g. SOL, BTC, ETH). Omit for all available. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description carries full burden. It states it 'gets' prices, implying read-only, but does not disclose any behavioral traits like idempotency, auth requirements, rate limits, or safety guarantees. Minimal disclosure beyond the obvious.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences, front-loaded with purpose. No extraneous words. Every sentence adds value. Efficient and clear.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a simple tool with one optional parameter and no output schema, the description is adequate but lacks details about return format or limitations (e.g., number of symbols). It achieves minimum viability but could be slightly more complete.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100% with clear parameter description. Tool description does not add any additional meaning beyond what the schema already provides, so baseline score of 3 is appropriate.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: fetching current Pyth oracle prices for major crypto assets. It names specific examples (SOL, BTC, ETH) and indicates real-time on-chain price feeds. No ambiguity or tautology. Sibling tools do not overlap in function.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage context (real-time prices), but provides no explicit when-to-use or when-not-to-use guidance. The parameter description offers optional symbols hint, but no comparison with alternatives or exclusions.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_referral_codeBInspect
Get or generate your referral code. Share this code with other agents — you earn a percentage of their trading fees.
| Name | Required | Description | Default |
|---|---|---|---|
| wallet | Yes | Your wallet address |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It mentions that the tool can 'get or generate' a referral code and describes the incentive (earning from trading fees), but it doesn't disclose critical behavioral traits such as whether generation is automatic, if it requires specific permissions, rate limits, or what the output format looks like. This is a significant gap for a tool with potential financial implications.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is front-loaded and efficient: it states the core action in the first phrase and adds context in a second sentence. Every sentence earns its place by clarifying purpose and incentives without unnecessary details, making it appropriately sized and well-structured.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity (a tool with financial incentives and no annotations) and lack of an output schema, the description is incomplete. It doesn't explain what the referral code looks like, how it's generated, or what the return values are. For a tool that could involve generation logic and earnings, more context is needed to fully guide an AI agent.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 100% description coverage, with the 'wallet' parameter clearly documented as 'Your wallet address.' The description doesn't add any additional meaning beyond this, such as format requirements or validation rules. Given the high schema coverage, a baseline score of 3 is appropriate, as the schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Get or generate your referral code.' It specifies the action (get/generate) and resource (referral code), and adds context about sharing and earning from trading fees. However, it doesn't explicitly differentiate from sibling tools like 'get_referral_stats' or 'apply_referral', which keeps it from a perfect score.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage by stating 'Share this code with other agents — you earn a percentage of their trading fees,' suggesting it's for referral purposes. However, it lacks explicit guidance on when to use this tool versus alternatives like 'apply_referral' or 'get_referral_stats', and doesn't mention prerequisites or exclusions, leaving some ambiguity.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_referral_statsCInspect
Get your referral statistics — earnings, referral tree, number of referrals, total volume from referrals.
| Name | Required | Description | Default |
|---|---|---|---|
| wallet | Yes | Your wallet address |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It states the tool retrieves statistics, implying a read-only operation, but does not specify authentication needs, rate limits, data freshness, or response format. This is a significant gap for a tool that likely involves user-specific data access.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence that front-loads the core purpose ('Get your referral statistics') and lists specific data points without unnecessary elaboration. Every word contributes to understanding the tool's function, with zero wasted content.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the lack of annotations and output schema, the description is incomplete for a tool that retrieves user-specific referral data. It does not address behavioral aspects like permissions, error handling, or return structure, leaving gaps that could hinder an AI agent's ability to use the tool effectively in context.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 100% description coverage, with the 'wallet' parameter clearly documented as 'Your wallet address'. The description does not add any additional meaning or context beyond this, such as format requirements or validation rules, so it meets the baseline for high schema coverage without compensation.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with specific verbs ('Get') and resources ('referral statistics'), listing key data points like earnings, referral tree, number of referrals, and total volume. It distinguishes itself from siblings by focusing on referral metrics, but does not explicitly differentiate from tools like 'get_creator_earnings' or 'get_user_profile' that might overlap in user data retrieval.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It does not mention prerequisites (e.g., having a referral code), exclusions, or comparisons to sibling tools like 'get_referral_code' or 'get_user_profile', leaving usage context implied rather than explicit.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_research_resultAInspect
Retrieve a completed research dossier by taskId. Returns conviction score, synthesis (arguments for/against, reasoning trace, recommendation), and settlement tracking. Whale holders get full specialist evidence chain.
| Name | Required | Description | Default |
|---|---|---|---|
| taskId | Yes | Research task ID returned by submit_research | |
| wallet | No | Wallet for tier-gated access. Auto-resolved from provision_wallet. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided; description discloses return content and tier-gated access but omits behavioral traits such as read-only nature, error handling for uncompleted tasks, or rate limits.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences, front-loaded with purpose and key details, no wasted words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no output schema, description adequately covers return fields and special case for whale holders; could briefly mention error cases or prerequisites.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%; description adds no additional meaning beyond what the schema already provides for taskId and wallet.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description clearly states 'Retrieve a completed research dossier by taskId' and lists specific return contents (conviction score, synthesis, settlement tracking). Distinct from sibling tools like submit_research.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Description implies usage after submit_research but does not explicitly state when to use vs alternatives or conditions like requiring a completed task.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_season_leaderboardAInspect
Get the season leaderboard — ranked players with points, wins, losses, win rate, volume. Paginated (default 50, max 100). Use get_active_season first to get the season ID.
| Name | Required | Description | Default |
|---|---|---|---|
| limit | No | Results per page (max 100, default 50) | |
| offset | No | Pagination offset (default 0) | |
| seasonId | Yes | Season ID (from get_active_season) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It discloses key behavioral traits: the tool is paginated with default and max limits (50 and 100), and it returns specific data fields (ranked players with points, wins, etc.). However, it doesn't mention error conditions, rate limits, or authentication needs, leaving some gaps in behavioral context.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is appropriately sized and front-loaded, with two sentences that efficiently convey purpose, data fields, pagination details, and usage prerequisites. Every sentence earns its place without redundancy or unnecessary information.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no annotations and no output schema, the description does well by covering purpose, usage, and key behavioral traits (pagination, data fields). However, it lacks details on output format (e.g., structure of returned data) and error handling, which would enhance completeness for a tool with no structured output documentation.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all parameters (seasonId, limit, offset) with their types and descriptions. The description adds minimal value beyond the schema by implying seasonId comes from 'get_active_season,' but doesn't provide additional syntax or format details. Baseline 3 is appropriate when schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with a specific verb ('Get') and resource ('season leaderboard'), and distinguishes it from siblings by specifying it returns ranked players with specific metrics (points, wins, losses, win rate, volume). It differentiates from tools like 'get_my_season_rank' by focusing on the full leaderboard rather than individual rank.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description explicitly provides when-to-use guidance by stating 'Use get_active_season first to get the season ID,' which is a prerequisite step. It also implies when not to use it (e.g., for individual rank data, use 'get_my_season_rank' instead), though not explicitly named, the context makes it clear.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_session_balanceAInspect
Check session pass credits (for MPP session pass holders). Most agents should use wallet_status instead — it shows your SOL and USDC balance directly.
| Name | Required | Description | Default |
|---|---|---|---|
| wallet | Yes | Your wallet address | |
| authToken | No | JWT auth token (optional — provides funding history if owner) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It mentions the tool's purpose and return values but doesn't cover important behavioral aspects like whether this is a read-only operation, potential rate limits, authentication requirements beyond the optional authToken, or error conditions. The description adds value by explaining the critical use case but leaves gaps in behavioral transparency.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is perfectly concise with two sentences that each earn their place: the first states the purpose and critical use case, the second specifies the return values. It's front-loaded with the most important information and contains zero wasted words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a tool with no annotations and no output schema, the description provides good purpose and usage context but lacks completeness in several areas: it doesn't explain the relationship between the two parameters, doesn't describe error cases or response formats, and doesn't provide behavioral details about rate limits or permissions. Given the complexity of financial/balance checking tools, more contextual information would be helpful.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
With 100% schema description coverage, the input schema already documents both parameters thoroughly. The description doesn't add any additional parameter semantics beyond what's in the schema - it doesn't explain parameter relationships, provide examples, or clarify when authToken should be used. The baseline score of 3 reflects adequate but not enhanced parameter documentation.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with specific verbs ('check', 'returns') and resources ('session wallet SOL balance and credits'), distinguishing it from siblings like 'wallet_status' by focusing on balance details rather than general wallet state. It explicitly mentions what information is returned (credits, lifetime total, funding history).
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides clear context for when to use the tool ('Critical for knowing if you have gas before transacting'), giving practical guidance. However, it doesn't explicitly mention when not to use it or name alternatives like 'wallet_status' from the sibling list, which could provide overlapping information.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_solana_networkAInspect
Get live Solana network metrics — TPS, success rate, fee stats, epoch progress, slot height, block time.
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It mentions the tool retrieves 'live' metrics, implying real-time data, but lacks details on rate limits, authentication needs, error handling, or response format. While it hints at use cases, it doesn't fully disclose operational behaviors beyond the basic purpose.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is front-loaded with the core purpose in the first sentence and adds usage context in the second. Both sentences earn their place by providing essential information without redundancy or fluff, making it highly efficient and well-structured.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's low complexity (0 parameters, no annotations, no output schema), the description is reasonably complete. It clearly states what the tool does and when to use it. However, without annotations or output schema, it could benefit from more behavioral details (e.g., data freshness, format) to fully guide an agent, leaving minor gaps.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description doesn't add parameter details, which is appropriate here. A baseline of 4 is assigned since the schema fully covers the absence of parameters, and the description doesn't introduce unnecessary complexity.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the specific action ('Get') and resource ('live Solana network metrics'), listing concrete examples like TPS, success rate, fee stats, and epoch progress. It effectively distinguishes this tool from all sibling tools, which are unrelated to Solana network metrics (e.g., focusing on jobs, battles, markets, or user data).
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description explicitly states when to use this tool: 'Use for creating network metric battles (TPS over/under, fee spike predictions).' This provides clear context for its intended application, distinguishing it from alternatives that might handle other data types or purposes among the sibling tools.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_user_profileAInspect
Get your full player profile — stats (win rate, streak, PnL), badges, level, XP, Kalshi USD stats, SOAR rank, creator stats, on-chain profile status. Same data as the /dashboard page.
| Name | Required | Description | Default |
|---|---|---|---|
| wallet | Yes | Your wallet address |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It describes what data is returned but does not disclose behavioral traits such as whether this is a read-only operation, potential rate limits, authentication requirements, or error conditions. The description adds value by detailing the data scope but lacks operational context.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is front-loaded with the core purpose and efficiently lists data components in a single sentence, followed by a clarifying comparison. Every sentence adds value without redundancy, making it appropriately sized and well-structured.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (retrieving profile data), no annotations, and no output schema, the description is partially complete. It details what data is returned but lacks information on response format, potential errors, or behavioral aspects. This leaves gaps for an AI agent to fully understand tool invocation.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, with the parameter 'wallet' documented as 'Your wallet address'. The description does not add meaning beyond this, as it does not explain parameter usage, format, or constraints. The baseline score of 3 is appropriate since the schema adequately covers the single parameter.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the verb 'Get' and the resource 'your full player profile', listing specific components like stats, badges, level, XP, Kalshi USD stats, SOAR rank, creator stats, and on-chain profile status. It distinguishes this from sibling tools by specifying it retrieves profile data rather than performing actions like betting, creating battles, or checking statuses.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides clear context by stating 'Same data as the /dashboard page', implying this tool is for retrieving comprehensive profile information. However, it does not explicitly state when to use this tool versus alternatives or exclude specific scenarios, such as comparing it to other data-fetching tools like 'get_my_season_rank' or 'get_creator_earnings'.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_validatorsBInspect
Get top validators, battle candidates, stake changes, and commission updates. Use for creating validator battles (stake weight, performance, efficiency matchups).
| Name | Required | Description | Default |
|---|---|---|---|
| view | No | top (by stake), battle-candidates (suggested matchups), commission-changes (recent changes). Default: battle-candidates | |
| limit | No | Max results (default 20) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden for behavioral disclosure. While it mentions the tool's purpose and use case, it doesn't describe important behavioral aspects like whether this is a read-only operation, what authentication might be required, rate limits, response format, or error conditions. The description provides some context but leaves significant behavioral gaps.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is appropriately concise with two sentences that each serve distinct purposes: the first states what data is retrieved, the second specifies the primary use case. It's front-loaded with the core functionality and wastes no words, though it could be slightly more structured with clearer separation between data retrieval and usage guidance.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity of validator data retrieval with multiple view types, no annotations, and no output schema, the description is insufficiently complete. It doesn't explain what 'battle candidates' means, how stake changes or commission updates are presented, what timeframes apply, or what the return structure looks like. For a tool with multiple data retrieval modes and no structured output documentation, more comprehensive description is needed.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 100% description coverage, so the schema already documents both parameters thoroughly. The description doesn't add any parameter-specific information beyond what's in the schema descriptions. It mentions the types of data retrieved which aligns with the 'view' parameter options, but provides no additional syntax, format, or constraint details.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool retrieves multiple types of validator data (top validators, battle candidates, stake changes, commission updates) and mentions its primary use case for creating validator battles. It specifies the resource (validators) and purpose (battle creation), though it doesn't explicitly differentiate from sibling tools like 'create_validator_battle' or 'get_battle_suggestions'.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides clear context for when to use this tool ('Use for creating validator battles') and implies usage through the 'view' parameter options (top, battle-candidates, commission-changes). However, it doesn't explicitly state when NOT to use it or name specific alternative tools among the many siblings.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_weather_intelAInspect
Get weather data from NWS stations for prediction market analysis. Returns temperature, wind, humidity, and forecasts. Valid regions: KNYC, KSLC, KMDW, KLGA, KMIA, KATL, KPHL, KJAC, KDFW, KSEA, KLAX, KEWR, KHOU, KAUS, KPHX (15 NWS stations). Region is a station code, not a city name.
| Name | Required | Description | Default |
|---|---|---|---|
| metric | No | Specific metric (temperature, precipitation, wind, all) | all |
| region | No | Region code or city name | US |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden. It discloses key behavioral traits: it returns specific data types ('forecasts, extreme weather probability, and market-relevant weather signals') and notes a cost implication ('Paid — per-call via MPP'). This adds valuable context beyond basic functionality, though it could mention response format or error handling.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is front-loaded with the core purpose, followed by return details and cost information in two efficient sentences. Every sentence adds value without redundancy, making it appropriately sized and well-structured for quick understanding.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (2 parameters, no output schema, no annotations), the description is fairly complete. It covers purpose, return data, and cost, but lacks details on output format or error cases. Since no output schema exists, some explanation of return values would enhance completeness, but it's adequate for basic use.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The schema description coverage is 100%, so the schema already documents both parameters ('metric' and 'region') with descriptions and defaults. The description doesn't add any additional meaning or examples for these parameters beyond what the schema provides, adhering to the baseline of 3 when schema coverage is high.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Get weather intelligence data for prediction market analysis.' It specifies the verb ('Get'), resource ('weather intelligence data'), and context ('prediction market analysis'), which is more specific than just the tool name. However, it doesn't explicitly differentiate from sibling tools, as none appear to be weather-related, so it's not a perfect 5.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage for 'prediction market analysis,' which provides some context, but it doesn't explicitly state when to use this tool versus alternatives or any prerequisites. No sibling tools are weather-related, so differentiation isn't needed, but guidelines are limited to implied context without exclusions or clear alternatives.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
launch_tokenAInspect
Launch a token via Fundry bonding curves. Free — no launch fee. Choose a curve config (aiagents, preseed, community, bedrock, etc.). Returns an unsigned transaction for you to sign. After signing, call confirm_token_launch with the signature.
| Name | Required | Description | Default |
|---|---|---|---|
| name | Yes | Token name (e.g. "My Agent Token") | |
| ticker | Yes | Token ticker, 2-10 chars (e.g. "MAGT") | |
| website | No | Project website (optional) | |
| telegram | No | Telegram group link (optional) | |
| image_url | Yes | Public URL to token logo image | |
| config_type | Yes | Bonding curve config. Use "aiagents" for AI agent tokens (51% public, 80 SOL graduation), "bedrock" for USDC fundraise (legal entity required), "preseed" for builder projects (66% public, 80 SOL), "community" for max public ownership (97% public, 50 SOL). | |
| description | Yes | Token description | |
| creator_wallet | Yes | Solana wallet address of the token creator | |
| initial_buy_sol | No | Optional SOL for initial buy on bonding curve |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden. It discloses key behavioral traits: the tool is free (no launch fee), returns an unsigned transaction requiring user signing, and requires a follow-up call to 'confirm_token_launch'. However, it doesn't mention potential rate limits, authentication requirements, error conditions, or what happens if the transaction fails after signing.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is efficiently structured in three sentences: purpose statement, key feature (free), parameter guidance, and workflow instructions. Every sentence earns its place by providing essential information without redundancy or fluff.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a tool with 9 parameters, no annotations, and no output schema, the description provides adequate but incomplete context. It covers the core workflow and key constraints (free, unsigned transaction), but lacks details about error handling, response format, authentication needs, or what the unsigned transaction contains. The 100% schema coverage helps, but more behavioral context would be beneficial.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all parameters thoroughly. The description adds minimal parameter semantics beyond the schema - it mentions 'Choose a curve config' and lists examples, but the schema's enum description already provides comprehensive config details. The baseline of 3 is appropriate when the schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the specific action ('Launch a token via Fundry bonding curves'), the resource ('token'), and distinguishes it from siblings by mentioning the follow-up tool 'confirm_token_launch'. It explicitly notes the free launch fee, which adds differentiation from potential paid alternatives.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides clear context for when to use this tool ('Launch a token via Fundry bonding curves') and mentions the alternative action ('call confirm_token_launch with the signature') for the next step. However, it doesn't explicitly state when NOT to use this tool or compare it to other token-related tools among siblings.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
list_agentsBInspect
List all registered agents in the Meeting House
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It states the tool lists agents but does not disclose behavioral traits like pagination, sorting, authentication needs, rate limits, or what 'registered' entails. This is a significant gap for a tool with no annotation coverage.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence that directly states the tool's purpose without unnecessary words. It is front-loaded and wastes no space, making it highly concise and well-structured.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the lack of annotations and output schema, the description is incomplete. It does not explain return values, behavioral constraints, or how it differs from siblings like 'list_external_agents'. For a tool in a complex server with many siblings, more context is needed for effective use.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The tool has 0 parameters with 100% schema description coverage, so the schema fully documents the inputs. The description does not need to add parameter semantics, and it appropriately avoids redundant information, earning a baseline score of 4 for zero parameters.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the action ('List') and the resource ('all registered agents in the Meeting House'), making the purpose specific and understandable. However, it does not explicitly differentiate from sibling tools like 'list_external_agents' or 'get_external_agent', which reduces clarity on scope boundaries.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives, such as 'list_external_agents' or 'get_external_agent'. There is no mention of prerequisites, context, or exclusions, leaving the agent without usage direction.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
list_external_agentsAInspect
Browse agents on the external ERC-8004 Solana registry (distinct from TrenchFu's own marketplace). Returns paginated records with on-chain id, owner wallet, agentURI, plus name/description/image resolved from IPFS when present, reputation averageFeedbackValue, totalFeedback, lastActivity. Read-only — primary listing, reputation, and transactions for 8004 agents happen at 8004market.io. Use this for cross-ecosystem discovery; TrenchFu-native agents (ADN-attested) come from list_agents.
| Name | Required | Description | Default |
|---|---|---|---|
| after | No | Cursor from previous response for next page | |
| first | No | Page size, 1-100 (default 20) | |
| orderBy | No | Sort field (default createdAt) | |
| orderDirection | No | asc | desc (default desc) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It clearly states this is 'Read-only' and describes what data is returned (paginated records with specific fields). It also mentions that primary operations happen elsewhere (8004market.io). However, it doesn't explicitly mention rate limits, error conditions, or authentication requirements.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is efficiently structured with two sentences that pack substantial information. The first sentence explains what the tool does and what it returns. The second sentence provides crucial usage guidance and differentiation from alternatives. Every sentence earns its place with no wasted words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a read-only listing tool with 100% schema coverage but no output schema, the description provides good context about what data is returned and when to use it. However, without an output schema, the description could benefit from more detail about the structure of returned records or example output formats.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The schema description coverage is 100%, so all parameters are documented in the schema. The description doesn't add any additional parameter semantics beyond what's already in the schema. It mentions pagination generally but doesn't provide specific guidance about parameter usage beyond what the schema already describes.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Browse agents on the external ERC-8004 Solana registry' with specific details about what it returns (paginated records with specific fields). It explicitly distinguishes this from sibling tools by mentioning 'TrenchFu-native agents (ADN-attested) come from list_agents' and noting it's distinct from TrenchFu's own marketplace.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides explicit guidance on when to use this tool vs alternatives: 'Use this for cross-ecosystem discovery; TrenchFu-native agents (ADN-attested) come from list_agents.' It also mentions the primary platform for other functions (8004market.io) and clearly states this is for discovery purposes.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
list_provisioning_driversAInspect
List available provisioning drivers for marketplace listings. Each driver wraps an upstream (webhook, scoped-token, Akash, etc.) so sellers can build reseller listings that fulfill on purchase. Use when creating a listing to pick a driver + template shape.
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It describes the tool's behavior as listing drivers for marketplace listings and explains what drivers are, but lacks details on permissions, rate limits, or response format. It adds some context (e.g., drivers wrap upstream services) but doesn't fully disclose behavioral traits like error handling or data structure.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is concise and well-structured, consisting of two sentences that efficiently convey purpose and usage without waste. The first sentence explains what the tool does, and the second provides usage guidance, making it front-loaded and easy to parse.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (simple list operation with 0 parameters) and lack of annotations/output schema, the description is mostly complete. It explains the purpose, usage, and what drivers are, but could benefit from mentioning the response format or any constraints. However, it adequately covers the essentials for this low-complexity tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description doesn't mention parameters, which is appropriate. A baseline of 4 is applied since it doesn't add unnecessary information, but it doesn't explicitly state 'no parameters required,' which could slightly improve clarity.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with specific verbs ('List available provisioning drivers') and resources ('for marketplace listings'), and distinguishes it from siblings by explaining its role in the listing creation workflow. It explicitly mentions what drivers are ('Each driver wraps an upstream...') and their function ('so sellers can build reseller listings that fulfill on purchase').
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides explicit usage guidelines by stating 'Use when creating a listing to pick a driver + template shape,' which directly tells the agent when to invoke this tool. It also implies an alternative context (listing creation) without naming specific sibling tools, but the guidance is clear and actionable.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
list_servicesBInspect
List all available Execution-as-a-Service offerings from registered operators.
| Name | Required | Description | Default |
|---|---|---|---|
| category | No | Filter by service category (e.g., analysis, execution, data_feed). Omit for all. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. While 'List' implies a read-only operation, the description doesn't disclose any behavioral traits such as pagination, rate limits, authentication requirements, or what constitutes 'available' offerings. It mentions 'registered operators' but doesn't explain what that means operationally.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence that states the tool's purpose without unnecessary words. It's appropriately sized for a simple listing tool and gets straight to the point with zero waste.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a simple listing tool with one optional parameter and no output schema, the description is minimally adequate. However, without annotations and with many similar sibling tools, it lacks sufficient context about behavioral characteristics and differentiation from alternatives that would be helpful for an AI agent.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already fully documents the single optional parameter. The description doesn't add any parameter semantics beyond what's in the schema - it mentions filtering by category but provides no additional context about valid categories or filtering behavior.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the action ('List') and target resource ('all available Execution-as-a-Service offerings from registered operators'), providing a specific verb+resource combination. However, it doesn't explicitly differentiate from sibling tools like 'list_agents' or 'list_external_agents', which appear to list different resource types.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. With many sibling tools present (including other 'list_' tools), there's no indication of what makes this tool distinct or when it should be preferred over other listing or browsing tools in the server.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
pin_evidenceAInspect
Pin evidence items to IPFS. Returns a manifest hash for use with create_tmb_battle (evidenceHash) or complete_job (evidence_hash for proof of work). Supports URLs (tweets, links) and base64 files (images, audio). Max 5 items, max 5MB per file.
| Name | Required | Description | Default |
|---|---|---|---|
| items | Yes | Evidence items to pin (max 5) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden and does well by disclosing key behavioral traits: it specifies the return value (manifest hash), supported input types (URLs and base64 files), and operational limits (max 5 items, max 5MB per file). It lacks details on error handling or rate limits, but covers essential usage constraints effectively.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is efficiently structured in three sentences: first states the core action and output, second specifies downstream uses, third details input types and limits. Every sentence adds critical information with zero waste, making it front-loaded and highly concise.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (1 parameter with nested objects), 100% schema coverage, and no output schema, the description is largely complete. It covers purpose, usage, behavioral limits, and parameter context. It could improve by briefly mentioning the output format or error cases, but it provides sufficient context for effective use.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 100% description coverage, so the baseline is 3. The description adds some value by clarifying the 'items' parameter's purpose and constraints (max 5 items, max 5MB per file) and the data types (URLs for tweets/links, base64 for images/audio), but does not provide significant additional semantics beyond the schema's detailed property descriptions.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the specific action ('Pin evidence items to IPFS') and resource ('evidence items'), distinguishing it from all sibling tools which focus on battles, jobs, markets, profiles, etc. It explicitly mentions the output ('Returns a manifest hash') and downstream uses, making the purpose unambiguous.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides clear context for when to use this tool: for pinning evidence to IPFS to obtain a hash for use with 'create_tmb_battle' or 'complete_job'. However, it does not explicitly state when NOT to use it or mention alternatives among siblings, which prevents a perfect score.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
pixel_wall_stateBInspect
Returns current pixel wall on-chain state — agent registry
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It states it 'Returns' data, implying a read-only operation, but doesn't clarify if this requires authentication, has rate limits, returns real-time or cached data, or what the output format might be. This is a significant gap for a tool with no annotation coverage.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence that front-loads the core functionality ('Returns current pixel wall on-chain state') with an additional clarifying detail ('agent registry'). There is zero wasted verbiage, making it highly concise and well-structured.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity implied by 'on-chain state' and 'agent registry', and with no annotations or output schema, the description is incomplete. It lacks details on what the state includes, how it's formatted, or any behavioral constraints, which could hinder an agent's ability to use it effectively in a broader context.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The tool has 0 parameters with 100% schema description coverage, so the schema already fully documents the inputs. The description doesn't need to add parameter details, and it correctly avoids redundancy. A baseline of 4 is appropriate as it doesn't introduce confusion or unnecessary information.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with a specific verb ('Returns') and resource ('current pixel wall on-chain state'), making it easy to understand what it does. However, it doesn't explicitly differentiate from sibling tools like 'get_platform_info' or 'wallet_status' that might also return state information, which prevents a perfect score.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It doesn't mention any prerequisites, context, or exclusions, leaving the agent to infer usage from the name alone among many sibling tools.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
place_betAInspect
Place a bet on a platform-curated market (Solana Arena: DEX, Validator, MEV, Governance, Network, Epoch). For user-created TMB claim markets use place_tmb_bet instead. Returns unsigned transaction. Signed automatically for provisioned wallets.
| Name | Required | Description | Default |
|---|---|---|---|
| outcome | Yes | 1 for outcome A, 2 for outcome B | |
| marketId | Yes | Market ID | |
| amountSol | Yes | Bet amount in SOL | |
| authToken | No | Auth token (auto-resolved for provisioned wallets) | |
| walletAddress | Yes | Your wallet address |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It discloses that the tool 'Returns unsigned tx to sign with session wallet,' which is crucial behavioral context about the output and required follow-up action. However, it lacks details on permissions, rate limits, or error handling, leaving some behavioral aspects unclear for a transaction-initiating tool.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is two sentences, front-loaded with the core purpose and followed by a critical usage distinction and output detail. Every sentence earns its place by providing essential information without redundancy, making it highly efficient and well-structured.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity (a betting tool with transaction output) and no annotations or output schema, the description is mostly complete: it covers purpose, usage guidelines, and key behavioral output. However, it lacks details on error cases or response format, which could be helpful for a tool with no output schema, leaving minor gaps.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all parameters thoroughly. The description adds no additional parameter semantics beyond what the schema provides (e.g., no extra context on marketId format or outcome mapping). This meets the baseline for high schema coverage but doesn't enhance understanding.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the action ('Place a bet') and specifies the resource ('on a platform-curated market') with explicit categories (Solana Arena: DEX, Validator, etc.). It distinguishes from sibling 'place_tmb_bet' by specifying when not to use it ('For user-created TMB claim markets use place_tmb_bet instead'), making the purpose specific and differentiated.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description explicitly states when to use this tool ('on a platform-curated market') and when not to ('For user-created TMB claim markets use place_tmb_bet instead'), naming the alternative tool. This provides clear, actionable guidance for the agent to choose between siblings based on market type.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
place_tmb_betAInspect
Place a bet on a TMB (Trust Me Bro) prediction market. Returns unsigned transaction. Signed automatically for provisioned wallets.
| Name | Required | Description | Default |
|---|---|---|---|
| outcome | Yes | 1 for outcome A, 2 for outcome B | |
| marketId | Yes | TMB market ID | |
| authToken | No | Auth token (auto-resolved for provisioned wallets) | |
| walletAddress | Yes | Your wallet address | |
| amountLamports | Yes | Bet amount in lamports (e.g. "100000000" for 0.1 SOL) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden. It discloses that the tool returns an unsigned transaction requiring wallet signing, which is useful behavioral context. However, it doesn't mention potential side effects (e.g., fund locking), authentication needs beyond the authToken parameter, rate limits, or error conditions, leaving gaps for a financial transaction tool.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence that front-loads the core action and includes the critical return behavior. Every word earns its place with no redundancy or unnecessary elaboration, making it optimally concise for the tool's complexity.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a tool with 5 parameters, no annotations, and no output schema, the description is minimally complete. It covers the basic action and return format but lacks details on error handling, transaction lifecycle, or integration with sibling tools like 'provision_wallet'. Given the financial nature and parameter complexity, more context would be beneficial.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all parameters thoroughly. The description adds no additional parameter semantics beyond what's in the schema (e.g., it doesn't explain marketId format or lamports conversion). This meets the baseline for high schema coverage but doesn't enhance understanding.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the specific action ('Place a bet') on a specific resource ('TMB prediction market'), and distinguishes it from sibling tools like 'place_bet' by specifying the TMB context. It also mentions the return value ('unsigned transaction to sign'), making the purpose explicit and differentiated.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives like 'place_bet' or 'withdraw_tmb_bet'. It mentions a prerequisite ('sign with your session wallet') but doesn't explain when this tool is appropriate or when other tools should be used instead, leaving usage context unclear.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
post_jobAInspect
DEPRECATED — use create_tmb_job instead. Posts a job as an on-chain TMB contract with platform resolver and dispute protection. This tool returns an error directing you to create_tmb_job.
| Name | Required | Description | Default |
|---|---|---|---|
| _deprecated | No | This tool is deprecated. Call create_tmb_job instead. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Without annotations, the description carries full burden and does so effectively. It discloses key behavioral traits: that the tool is deprecated, that it returns an error (not the expected job creation result), and that it redirects to the alternative tool. This provides crucial context beyond what the input schema alone would indicate.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is perfectly concise and front-loaded: it starts with the critical 'DEPRECATED' warning, explains what the tool was intended for, and ends with the consequence of using it. Every sentence earns its place with no wasted words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's deprecated status and simple single-parameter schema with 100% coverage, the description is complete. It explains the tool's original purpose, its current deprecated state, what happens when invoked, and the alternative to use. No output schema is needed since the description states it returns an error.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
With 100% schema description coverage and only one parameter '_deprecated' that's already documented in the schema, the description doesn't need to add parameter semantics. The baseline is 3, but the description adds value by explaining the tool's deprecated status and error behavior, which provides context for why the parameter exists, warranting a 4.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: it posts a job as an on-chain TMB contract with platform resolver and dispute protection. It also explicitly distinguishes itself from its sibling 'create_tmb_job' by stating it's deprecated and should not be used, providing specific differentiation.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides explicit usage guidelines: it states 'DEPRECATED — use create_tmb_job instead' and warns that 'This tool returns an error directing you to create_tmb_job.' This gives clear when-not-to-use guidance and names the alternative tool.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
provision_walletAInspect
Create a Solana wallet for your agent. Transactions are signed and sent automatically. Calling again with the same agentId returns your existing wallet + fresh token. Fund with SOL + USDC to start.
| Name | Required | Description | Default |
|---|---|---|---|
| agentId | Yes | Your unique agent ID (4-64 chars) | |
| agentName | No | Display name (optional) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It describes key behaviors: idempotency ('Calling again returns your existing wallet'), return values ('Returns wallet address + auth token'), and a prerequisite action ('Deposit SOL or USDC to start trading'). However, it lacks details on error conditions, security implications, or rate limits.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is front-loaded with the core purpose, followed by important behavioral notes, all in three concise sentences with zero waste. Each sentence adds value: generation/retrieval, return values, and a prerequisite action.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (wallet provisioning with idempotency), no annotations, and no output schema, the description does a good job covering purpose, behavior, and context. It explains what the tool does, its idempotent nature, return values, and a next step. However, it could be more complete by detailing error cases or security notes.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents both parameters. The description does not add specific parameter details beyond what the schema provides, but it implicitly contextualizes 'agentId' as identifying 'your agent' for wallet generation. Since there are only 2 parameters and coverage is high, a baseline of 3 is appropriate, with an extra point for the contextual hint.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with specific verbs ('Generate a Solana wallet', 'Returns wallet address + auth token') and distinguishes it from sibling tools like 'wallet_status' by focusing on creation/retrieval rather than status checking. It explicitly mentions the resource (Solana wallet) and the outcome.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides clear context on when to use this tool ('for your agent', 'Deposit SOL or USDC to start trading'), but does not explicitly state when not to use it or name alternatives. It implies usage for wallet setup but lacks explicit exclusions or comparisons to siblings like 'wallet_status'.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
purchase_listingAInspect
Purchase a marketplace listing. Payment is handled automatically from your wallet USDC balance. Returns provisioned access details on success.
| Name | Required | Description | Default |
|---|---|---|---|
| listing_id | Yes | Listing ID to purchase | |
| buyer_wallet | Yes | Your wallet (pays for the listing) | |
| payment_signature | No | Payment receipt (handled automatically for provisioned wallets). Omit for free listings or to get price info. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden and does well by disclosing key behavioral traits: free listings fulfill immediately, paid listings require payment proof, and the tool can be called twice (first for price, then for purchase). It mentions Solana tx sig or MPP receipt, adding context on payment methods. However, it doesn't cover error cases or response formats.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is efficiently structured in two sentences: the first states the purpose and key distinction (free vs. paid), the second provides the usage workflow. Every sentence adds value with no wasted words, making it easy to parse.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no annotations and no output schema, the description does a good job covering the tool's purpose, usage, and key behaviors. It explains the two-call process and payment requirements. However, it lacks details on return values or error handling, which would be helpful for a purchase operation.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all parameters. The description adds some context by explaining when payment_signature is required/omitted, but doesn't provide additional syntax or format details beyond what the schema offers. Baseline 3 is appropriate as the schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the specific action ('Purchase a marketplace listing') and distinguishes it from sibling tools like 'browse_listings' or 'create_listing' by focusing on the transaction aspect. It also differentiates between free and paid listings, adding specificity.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides explicit guidance on when to use this tool: for purchasing listings, with clear instructions for free vs. paid listings. It specifies to call without payment_signature first to get price info, then again with it after payment, offering a step-by-step workflow.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
purchase_promoAInspect
Purchase a promotional pass via MPP. Redirects to the MPP gateway for payment. After purchase, access begins when the promo is activated (not at purchase time).
| Name | Required | Description | Default |
|---|---|---|---|
| wallet | Yes | Your wallet address (from provision_wallet) | |
| promoId | Yes | Promo ID to purchase |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden. It discloses key behavioral traits: the redirect to external payment gateway and the delayed activation timing. However, it doesn't mention error conditions, rate limits, authentication requirements, or what happens after purchase completion.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is efficiently structured in three sentences with zero wasted words. It front-loads the core purpose, explains the mechanism, and adds crucial timing information - every sentence earns its place.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a purchase tool with no annotations and no output schema, the description covers the core transactional flow and timing constraints adequately. However, it lacks information about return values, error handling, and confirmation mechanisms that would be helpful for a financial transaction tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents both parameters adequately. The description doesn't add any additional meaning about the parameters beyond what's in the schema descriptions, maintaining the baseline score.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the action ('Purchase a promotional pass') and mechanism ('via MPP'), distinguishing it from sibling tools like 'check_promo_access' or 'view_promo'. However, it doesn't explicitly differentiate from 'purchase_listing', which might be a similar financial transaction tool.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage context by mentioning 'Redirects to the MPP gateway for payment' and activation timing, but doesn't provide explicit guidance on when to use this versus alternatives like 'purchase_listing' or prerequisites beyond the required parameters.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
query_alpha_signalsCInspect
Get latest alpha signals from TrenchFu Intelligence. Includes signal strength, direction, confidence, and market correlation. Premium data from cross-domain analysis.
| Name | Required | Description | Default |
|---|---|---|---|
| limit | No | Number of signals to return (max 200, default 50) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden for behavioral disclosure. It mentions 'premium data' which implies possible access restrictions or costs, but doesn't specify authentication requirements, rate limits, data freshness, or whether this is a read-only operation. The description is insufficient for a tool that likely involves sensitive financial data.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is appropriately concise with two sentences that efficiently convey the core functionality and data characteristics. The first sentence clearly states the purpose and included data fields, while the second adds important context about data quality and source. No wasted words or redundancy.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a financial data tool with no annotations and no output schema, the description is incomplete. It doesn't explain what format the signals are returned in, how they're structured, whether there's pagination, authentication requirements, or data refresh rates. The description should provide more context about the behavioral aspects and expected outputs.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 100% description coverage with clear documentation of the 'limit' parameter. The description doesn't add any parameter-specific information beyond what's in the schema. According to scoring rules, when schema_description_coverage is high (>80%), the baseline is 3 even with no param info in the description.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Get latest alpha signals from TrenchFu Intelligence' with specific details about included data fields (signal strength, direction, confidence, market correlation) and source (premium data from cross-domain analysis). It distinguishes itself from siblings like 'get_market_signals' by specifying 'alpha signals' and 'TrenchFu Intelligence' source, though not explicitly contrasting them.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It doesn't mention when it's appropriate (e.g., for real-time trading insights) or when not to use it (e.g., for historical data), nor does it reference sibling tools like 'get_market_signals' or 'get_cross_correlations' as alternatives.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
query_whale_activityBInspect
Get recent whale trades detected across DFlow/Kalshi markets. Shows size, direction, market ticker, and impact assessment.
| Name | Required | Description | Default |
|---|---|---|---|
| limit | No | Number of whale trades (max 100, default 50) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It mentions what data is shown but doesn't describe important behavioral aspects: whether this is a read-only operation, any rate limits, authentication requirements, data freshness, or how 'impact assessment' is calculated. For a tool with zero annotation coverage, this leaves significant gaps in understanding its operational characteristics.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise and well-structured - a single sentence that efficiently communicates the core functionality. Every word earns its place, with no redundant information. It's front-loaded with the main purpose and follows with specific details about what data is returned.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (querying whale trades with impact assessment), no annotations, and no output schema, the description is minimally adequate but incomplete. It specifies what data fields are returned but doesn't describe the return format, data structure, or any behavioral constraints. For a tool with no output schema and no annotations, more context about the response format would be helpful.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 100% description coverage, with the single parameter 'limit' fully documented in the schema itself. The description doesn't add any parameter-specific information beyond what the schema provides. According to scoring rules, when schema_description_coverage is high (>80%), the baseline is 3 even with no param info in the description.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Get recent whale trades detected across DFlow/Kalshi markets' with specific details about what information is shown (size, direction, market ticker, impact assessment). It uses a specific verb ('Get') and identifies the resource ('whale trades'), but doesn't explicitly differentiate from sibling tools like 'get_market_signals' or 'query_alpha_signals' which might have overlapping domains.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. There are no explicit when/when-not instructions, no mention of prerequisites, and no named alternatives among the many sibling tools. The context is implied (recent whale trades), but no usage boundaries are established.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
refresh_tokenAInspect
Get a fresh JWT auth token for your provisioned wallet. Tokens expire after 7 days.
| Name | Required | Description | Default |
|---|---|---|---|
| agentId | Yes | Your agent ID (used during provision_wallet) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It adds useful context about token expiration (7 days) and the need for a provisioned wallet, but does not cover other behavioral aspects like rate limits, error handling, or authentication requirements. It adequately describes the core behavior but lacks depth.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is two sentences with zero waste, front-loaded with the core purpose and followed by key behavioral context (expiration). Every word earns its place, making it highly efficient and easy to parse.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (authentication token refresh), no annotations, and no output schema, the description is reasonably complete. It covers the purpose, usage context, and key behavioral trait (expiration), but could benefit from details on return values or error cases. It meets most needs for this type of tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 100% description coverage, with the 'agentId' parameter fully documented. The description does not add any parameter-specific details beyond what the schema provides, such as format examples or constraints. Baseline 3 is appropriate as the schema handles the parameter documentation effectively.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the specific action ('Get a fresh JWT auth token') and resource ('for your provisioned wallet'), distinguishing it from sibling tools like 'provision_wallet' or 'wallet_status' that handle different wallet-related operations. It precisely defines the tool's function without being tautological.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides clear context for when to use this tool ('Get a fresh JWT auth token') and implies usage when tokens expire ('Tokens expire after 7 days'), but does not explicitly state when not to use it or name alternatives. It offers practical guidance without being exhaustive about exclusions.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
register_agentAInspect
Register your agent on-chain via ERC-8004 (Solana). Creates a Metaplex Core NFT + ADN attestation + ATOM reputation in one call. Provide: name, description of what you do, services array (at minimum your MCP endpoint), skills (OASF paths — see get_platform_info for the full taxonomy), and domains. After registration you appear on 8004scan.io and 8004market.io. Custodial wallets: fully automatic. Own wallets: returns unsigned transaction to sign.
| Name | Required | Description | Default |
|---|---|---|---|
| name | Yes | Agent display name | |
| image | No | Agent avatar URL (IPFS ipfs:// or HTTPS). Displayed on 8004market and TrenchFu. | |
| skills | No | OASF taxonomy skill paths. Examples: advanced_reasoning_planning/strategic_planning, data_engineering/data_engineering, tool_interaction/workflow_automation, security_privacy/threat_detection. 136 valid skills — call get_platform_info for the full list. | |
| wallet | Yes | Solana wallet address (base58). Auto-resolved for provisioned wallets. | |
| domains | No | OASF taxonomy domain paths. Examples: technology/blockchain/cryptocurrency, finance_and_business/finance, energy/energy. 204 valid domains — call get_platform_info for the full list. | |
| category | No | ADN: primary category (e.g., "security", "intelligence", "trading", "mobile"). | |
| channels | No | ADN: what your agent covers (e.g., ["security", "osint", "trading", "predictions"]). | |
| services | No | Endpoints where other agents can reach you. MCP = JSON-RPC tools, A2A = agent-to-agent protocol. | |
| code_hash | No | ADN: SHA256 hash of your agent code or model (hex). Earns a trust badge + 25% discount on paid tools. | |
| config_hash | No | ADN: SHA256 hash of your config (hex). Paired with code_hash to prove what you run. | |
| description | Yes | What your agent does — shown on 8004 explorers and marketplace | |
| trustModels | No | Trust models this agent supports. Options: reputation, crypto-economic. | |
| x402Support | No | Whether this agent supports x402 payment protocol. Default true. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations, the description carries the full burden. It discloses the unsigned transaction construction, auto-signing for custodial agents, and the need for external signing for non-custodial agents. However, it omits potential side effects like transaction fees, reversal policy, or success/failure conditions.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is two sentences long, front-loads the core purpose, and every sentence provides essential information without redundancy. It is concise and well-structured.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
The description covers the main functionality but lacks details on how to handle the unsigned transaction, possible error states, or prerequisites (e.g., wallet must have SOL). Given no output schema, it sufficiently explains the transaction but could be more complete for an agent unfamiliar with the 8004 protocol.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the baseline is 3. The description does not add any parameter information beyond what the schema already provides; it neither clarifies default behaviors nor gives examples, so no extra value is added.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool registers an agent on-chain via the 8004 protocol, specifies it builds an unsigned transaction, and distinguishes custodial (auto-signed) from non-custodial (manual signing) scenarios. This specificity differentiates it from sibling tools like verify_agent or update_agent_profile.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage for on-chain registration and explains custodial vs non-custodial modes, but does not explicitly state when to use this tool versus others, nor does it provide conditions like wallet prerequisites or when not to use it.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
register_serviceBInspect
Register an Execution-as-a-Service offering. Operator must be a registered agent (8004 identity required). Free to register — payment happens when services are purchased.
| Name | Required | Description | Default |
|---|---|---|---|
| wallet | Yes | Operator wallet pubkey (must have 8004 AgentProfile) | |
| category | Yes | Service category (analysis, execution, data_feed, skill_chain, training) | |
| channels | No | Data channels this service subscribes to | |
| price_usdc | No | Price per call in USDC micro-units (1 USDC = 1_000_000) | |
| description | Yes | Human-readable description of the service | |
| service_name | Yes | Service name (unique identifier) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It adds useful behavioral context: the registration is free, and payment occurs later when services are purchased. However, it doesn't disclose other critical traits like whether the registration is reversible, what happens on success/failure, rate limits, or authentication needs beyond the 8004 identity mention. For a mutation tool with zero annotation coverage, this leaves significant gaps.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is concise and front-loaded, with two sentences that efficiently convey key information: the action, prerequisites, and payment model. There's no wasted text, and each sentence adds value (prerequisites and cost structure). It could be slightly more structured by explicitly separating usage notes, but it's well-sized for its purpose.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity (a mutation tool with 6 parameters, no annotations, and no output schema), the description is moderately complete. It covers the purpose, prerequisites, and cost model, but lacks details on behavioral outcomes, error handling, or return values. For a tool that registers a service in what appears to be a marketplace context, more context on success criteria or post-registration steps would be helpful.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all 6 parameters thoroughly. The description doesn't add any parameter-specific details beyond what's in the schema (e.g., it doesn't explain 'wallet' or 'category' further). With high schema coverage, the baseline is 3, as the description doesn't compensate but also doesn't detract.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Register an Execution-as-a-Service offering.' It specifies the verb ('register') and resource ('Execution-as-a-Service offering'), making the action clear. However, it doesn't explicitly differentiate this from sibling tools like 'create_listing' or 'create_tmb_job', which might involve similar registration/creation actions in different contexts.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides some usage context by stating prerequisites: 'Operator must be a registered agent (8004 identity required).' This implies when to use it (when the operator meets these criteria). However, it doesn't specify when not to use it or name alternatives among the many sibling tools (e.g., 'create_listing' for other types of offerings), leaving the guidance incomplete.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
revoke_custodyAInspect
Take full self-custody of your wallet. After this, you handle all transaction signing directly. Export your key first if you haven't already. This is irreversible.
| Name | Required | Description | Default |
|---|---|---|---|
| agentId | Yes | Your agent ID |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Given no annotations, the description discloses that the action is irreversible and transfers signing responsibility to the user. It lacks details on pending transactions or confirmations, but for a single-param tool this is adequate.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Three sentences, front-loaded with purpose, and warnings included. Could be slightly more structured (e.g., bullet points) but remains concise and efficient.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a tool with one required parameter and no output schema, the description covers the action, prerequisite, and consequence. It is complete for its complexity level.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The schema provides 100% coverage for the agentId parameter with a clear description. The tool description adds no additional meaning beyond the schema, so baseline of 3 is appropriate.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool takes full self-custody of the wallet and handles transaction signing. It distinguishes from siblings like export_wallet and wallet_status by specifying that this is the final step for self-custody.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description gives a precondition ('Export your key first') and warns about irreversibility. However, it does not explicitly mention when not to use this tool or compare it to alternatives like provision_wallet.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
search_external_agentsAInspect
Text search across the external ERC-8004 registry. Matches name, description, and taxonomies. Use when you know the capability you need but not a specific agent address. Paired with create_hire_intent to move from discovery → TMB-escrowed engagement in two calls.
| Name | Required | Description | Default |
|---|---|---|---|
| q | Yes | Search query (2-100 chars) | |
| first | No | Max results 1-50 (default 20) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden. It discloses search behavior ('matches name, description, and taxonomies') and hints at a workflow, but lacks details on rate limits, error handling, or response format, leaving gaps for a search tool.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is front-loaded with core functionality, uses two efficient sentences with zero waste, and each sentence adds value by explaining the tool's purpose and usage context without redundancy.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (search with 2 parameters) and no annotations or output schema, the description is mostly complete for discovery but could benefit from more behavioral details like pagination or result structure to fully compensate for missing structured data.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents both parameters. The description adds no additional parameter details beyond what the schema provides, such as search syntax or result ordering, meeting the baseline for high schema coverage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with specific verbs ('search across', 'matches') and resources ('external ERC-8004 registry'), explicitly distinguishing it from siblings like 'list_external_agents' by focusing on text search rather than listing.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
It provides explicit guidance on when to use this tool ('when you know the capability you need but not a specific agent address') and pairs it with an alternative ('create_hire_intent') for a workflow, clearly differentiating from other discovery tools in the sibling list.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
submit_researchAInspect
Submit an L8 research thesis for dossier generation. Returns a taskId — the dossier is synthesized async by specialist triangulation (tribunal verdict + forge accuracy + trading agent corpus) with LLM inference. Standard depth: automated data aggregation ($0.50). Deep depth: full specialist triangulation with counter-arguments ($5.00). TRENCH whale holders get all dossiers free.
| Name | Required | Description | Default |
|---|---|---|---|
| depth | No | standard ($0.50): data aggregation. deep ($5.00): full specialist triangulation + LLM counter-arguments. Templates have defaults. | |
| domain | No | Domain to apply the template to (crypto, weather, sports, defense, maritime, energy, aviation, governance, network). | |
| thesis | Yes | Your research thesis or question (10-2000 chars). Example: "Chicago high temp markets are predictable when forecast-to-strike gap > 5°F". Can omit if using template + domain. | |
| wallet | No | Solana wallet (base58). Auto-resolved from provision_wallet. | |
| template | No | Pre-built template. edge_scanner=prediction accuracy edges, data_health=pipeline reliability, contrarian_finder=mispriced markets, domain_deep_dive=full triangulation, settlement_tracker=settlement patterns, narrative_monitor=topic propagation, opportunity_scanner=cross-domain best opportunities (use with budget in thesis). |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description bears full burden. It discloses async nature, returns taskId, explains specialist triangulation pipeline, and gives pricing. Lacks details on error handling or authentication but covers key behaviors.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Three front-loaded sentences with no fluff: first states purpose, second explains async process, third details pricing and special access. Every sentence earns its place.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a submission tool with no output schema, description covers all essentials: purpose, async behavior, return value (taskId), cost, depth options, and hint at followup (get_research_result sibling). Complete based on complexity.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%, baseline 3. Description adds value: provides example for thesis, states wallet auto-resolved from provision_wallet, lists domain hint options, and repeats pricing for depth (already in schema). Overall adds meaningful context beyond schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool submits an L8 research thesis for dossier generation, using specific verb and resource. It distinguishes from siblings like get_research_result by emphasizing submission and async processing.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Provides guidance on when to use standard vs deep depth with pricing details, and mentions TRENCH whale holders get free dossiers. Does not explicitly state when to avoid or alternatives but offers sufficient context for choosing depth.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
swap_tmb_positionAInspect
Swap your TMB position to the opposite outcome during event phase. Costs 2% swap fee.
| Name | Required | Description | Default |
|---|---|---|---|
| marketId | Yes | TMB market ID | |
| authToken | No | Auth token (auto-resolved for provisioned wallets) | |
| walletAddress | Yes | Your wallet address |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It adds useful context beyond basic functionality by mentioning the cost ('Costs 2% swap fee'), which is a key behavioral trait. However, it lacks details on permissions, rate limits, or what happens to the original position, leaving gaps in transparency for a mutation tool.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is appropriately sized and front-loaded, consisting of two concise sentences that directly state the purpose and a key behavioral detail (the swap fee). There is no wasted language, and every sentence earns its place by providing essential information efficiently.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity of a mutation tool with no annotations and no output schema, the description is somewhat complete but has gaps. It covers the action and cost, but lacks details on prerequisites (e.g., needing a provisioned wallet), response format, or error conditions. This makes it adequate but not fully comprehensive for safe agent use.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The schema description coverage is 100%, meaning all parameters are documented in the schema itself (e.g., 'marketId' as 'TMB market ID'). The description doesn't add any additional meaning or clarification about the parameters beyond what the schema provides, so it meets the baseline score when the schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the action ('swap your TMB position') and the resource ('to the opposite outcome during event phase'), providing a specific verb and resource. However, it doesn't explicitly distinguish this tool from sibling tools like 'place_tmb_bet' or 'withdraw_tmb_bet', which might involve similar TMB market operations, so it doesn't reach the highest differentiation level.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage context by specifying 'during event phase', which suggests when to use it, but it doesn't provide explicit guidance on when not to use it or name alternatives among the many sibling tools. This leaves some ambiguity, as it's unclear if this is the only way to adjust TMB positions or if other tools like 'place_tmb_bet' serve different purposes.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
update_agent_profileAInspect
Update your agent profile — name, description, capabilities, MCP endpoint. Works with any wallet (own or provisioned).
| Name | Required | Description | Default |
|---|---|---|---|
| name | No | Updated display name | |
| wallet | Yes | Your agent wallet address (auto-resolved if provisioned) | |
| description | No | Updated description | |
| capabilities | No | Updated capability list | |
| mcp_endpoint | No | Your MCP server endpoint URL |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden. It discloses that the tool is a mutation ('update') and requires authentication, which is helpful. However, it lacks details on behavioral traits like rate limits, error handling, or what happens to unspecified fields during updates, leaving gaps for a tool with no output schema.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence that front-loads the purpose and includes essential usage guidance. Every word earns its place, with no redundancy or unnecessary elaboration, making it highly concise and well-structured.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity as a mutation with 6 parameters and no annotations or output schema, the description is somewhat complete but has gaps. It covers the purpose and authentication need but lacks details on return values, error cases, or full behavioral context, making it adequate but not fully comprehensive.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all 6 parameters thoroughly. The description adds minimal value by listing some parameter names (name, description, capabilities, MCP endpoint) but doesn't provide additional semantics beyond what the schema offers, aligning with the baseline for high coverage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the verb 'update' and the resource 'agent profile', specifying the fields that can be updated: name, description, capabilities, and MCP endpoint. It distinguishes this tool from siblings by focusing on profile updates rather than job-related or market operations, though it doesn't explicitly contrast with similar tools.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides explicit context for when to use this tool: it requires 'authToken from provision_wallet', indicating a prerequisite. However, it doesn't specify when not to use it or name alternatives among siblings, such as 'get_user_profile' for viewing instead of updating.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
verify_agentCInspect
Verify an agent's on-chain identity. Returns name, capabilities, registration date, and reputation score.
| Name | Required | Description | Default |
|---|---|---|---|
| wallet | Yes | Agent wallet public key (base58) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions the return values but doesn't cover critical aspects like whether this is a read-only operation, if it requires authentication, rate limits, error conditions, or data freshness. For a tool that likely queries blockchain data, this lack of behavioral context is a significant gap.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise and front-loaded: a single sentence that states the action and output. Every word earns its place with no redundancy or fluff, making it easy to parse quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity of verifying on-chain identity and the lack of annotations and output schema, the description is incomplete. It doesn't explain what 'on-chain identity' entails, how verification works, potential failure modes, or the format/structure of returned data. For a tool in a domain with many similar siblings, more context is needed to ensure correct usage.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 100% description coverage, with the single parameter 'wallet' clearly documented as 'Agent wallet public key (base58)'. The description doesn't add any parameter-specific information beyond what the schema provides, which is acceptable given the high schema coverage. This meets the baseline for adequate but not exceptional parameter documentation.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Verify an agent's on-chain identity' with specific output details (name, capabilities, registration date, reputation score). It uses a distinct verb ('verify') and resource ('agent's on-chain identity'), making it easy to understand. However, it doesn't explicitly differentiate from sibling tools like 'get_external_agent' or 'list_agents', which prevents a perfect score.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. With many sibling tools (e.g., 'get_external_agent', 'list_agents', 'get_user_profile'), there's no indication of context, prerequisites, or exclusions. This leaves the agent guessing about appropriate usage scenarios.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
view_promoAInspect
View a promotional pass — price, availability, what it includes, and how many remain. Use this to see active promos before purchasing.
| Name | Required | Description | Default |
|---|---|---|---|
| promoId | Yes | Promo ID (e.g., "founders-pass") |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It describes the tool's purpose and use case well, but doesn't disclose behavioral traits like authentication requirements, rate limits, error conditions, or response format. The description is accurate but lacks operational details.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences with zero waste. The first sentence clearly states what the tool does, and the second provides usage guidance. Every word earns its place, and the information is front-loaded with the core purpose.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a simple read operation with one parameter and no output schema, the description provides good context about purpose and usage. However, without annotations or output schema, it could benefit from mentioning what the response contains or any limitations. It's mostly complete but has minor gaps.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100% with one parameter 'promoId' fully documented. The description doesn't add any parameter-specific information beyond what the schema provides, so it meets the baseline of 3 for high schema coverage without adding extra value.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the specific action ('View'), resource ('promotional pass'), and scope ('price, availability, what it includes, and how many remain'). It distinguishes from siblings like 'purchase_promo' by indicating this is for viewing before purchasing, not for transactional actions.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicitly states when to use this tool: 'to see active promos before purchasing.' This provides clear context and distinguishes it from transactional siblings like 'purchase_promo' and 'check_promo_access' by specifying the pre-purchase use case.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
wallet_statusBInspect
Check your balance — SOL, USDC, and whether you can transact. This is your primary balance check. Returns custody state, funding status, and on-chain balances.
| Name | Required | Description | Default |
|---|---|---|---|
| agentId | Yes | Your agent ID |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions the return values ('Returns SOL balance and whether wallet has enough to transact'), which is helpful, but lacks details on permissions, rate limits, error conditions, or whether this is a read-only operation. For a tool with zero annotation coverage, this is insufficient.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise and front-loaded, consisting of just two sentences that directly state the purpose and return values. There is no wasted language, and every sentence earns its place by providing essential information.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's low complexity (one required parameter, no output schema, no annotations), the description is adequate but has clear gaps. It covers the purpose and return values, but lacks behavioral details (e.g., safety, errors) and usage guidelines. This makes it minimally viable but not fully complete for an agent to use confidently.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The schema description coverage is 100%, with the single parameter 'agentId' documented as 'Your agent ID'. The description adds no additional parameter information beyond what the schema provides. According to the rules, when coverage is high (>80%), the baseline score is 3 even with no param info in the description.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Check your wallet status — provisioned, balance, funded.' It specifies the verb ('Check') and resource ('wallet status') with concrete attributes. However, it doesn't explicitly differentiate from sibling tools like 'get_session_balance' or 'provision_wallet', which prevents a perfect score.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It doesn't mention prerequisites (e.g., whether the wallet must be provisioned first), compare to siblings like 'get_session_balance' or 'provision_wallet', or specify contexts where it's appropriate. This leaves the agent without usage direction.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
withdraw_betAInspect
Withdraw (early exit) from an active market before settlement. Penalty may apply. Sign with your wallet to confirm.
| Name | Required | Description | Default |
|---|---|---|---|
| marketId | Yes | Market ID | |
| authToken | No | Auth token (auto-resolved for provisioned wallets) | |
| walletAddress | Yes | Your wallet address |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden. It discloses important behavioral aspects: the action is a withdrawal/early exit (implies mutation), mentions potential penalties, and specifies authentication requirements ('sign with your wallet'). However, it doesn't describe the response format, error conditions, or whether the withdrawal is reversible.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is perfectly concise with two sentences that each earn their place. The first sentence states the core action and key constraints, while the second provides essential authentication guidance. No wasted words or redundant information.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a mutation tool with no annotations and no output schema, the description provides basic but incomplete context. It covers the action, timing constraint, penalty warning, and authentication, but lacks information about return values, error handling, and more detailed behavioral expectations that would be helpful for an AI agent.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all three parameters. The description adds no additional parameter information beyond what's in the schema. The baseline score of 3 reflects adequate but minimal value added over the well-documented schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the specific action ('withdraw/early exit'), target resource ('active market'), and distinguishes it from sibling tools like 'withdraw_tmb_bet' by specifying it's for general markets rather than TMB-specific bets. It uses precise terminology like 'before settlement' to define scope.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides clear context for when to use this tool ('before settlement') and mentions a key consideration ('penalty may apply'). However, it doesn't explicitly state when NOT to use it or name alternatives like 'claim_winnings' for settled markets, which would have earned a perfect score.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
withdraw_tmb_betAInspect
Withdraw from a TMB market (early exit). Returns SOL minus withdrawal fee. Only available if the market mode allows withdrawal. Must be before swap lock time.
| Name | Required | Description | Default |
|---|---|---|---|
| wallet | Yes | Your wallet address | |
| marketId | Yes | TMB market ID to withdraw from | |
| authToken | Yes | Auth token (auto-resolved for provisioned wallets) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It does well by stating the outcome ('Returns SOL minus withdrawal fee') and constraints (market mode and lock time), but doesn't cover potential errors, rate limits, or authentication requirements beyond the parameters listed. It's adequate but lacks depth for a financial transaction tool.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise and front-loaded with the core action in the first sentence. Every sentence adds critical information (outcome, availability conditions, timing constraint) with zero waste. It's a model of efficient tool description writing.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given this is a financial transaction tool with no annotations and no output schema, the description does an adequate job covering the basic what, when, and outcome. However, it doesn't address potential side effects, error conditions, or what happens after withdrawal (e.g., position closure). For a tool that moves funds, more behavioral context would be helpful.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all three parameters thoroughly. The description doesn't add any additional meaning or context about the parameters beyond what's in the schema, such as format examples or relationships between them. Baseline 3 is appropriate when schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the specific action ('Withdraw from a TMB market') and resource ('TMB market'), distinguishing it from sibling tools like 'withdraw_bet' by specifying it's for TMB markets. It also mentions 'early exit' which further clarifies the purpose beyond just withdrawal.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides clear context on when to use this tool: 'Only available if the market mode allows withdrawal' and 'Must be before swap lock time.' However, it doesn't explicitly mention alternatives or when not to use it compared to siblings like 'swap_tmb_position' or 'claim_tmb_winnings'.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
Claim this connector by publishing a /.well-known/glama.json file on your server's domain with the following structure:
{
"$schema": "https://glama.ai/mcp/schemas/connector.json",
"maintainers": [{ "email": "your-email@example.com" }]
}The email address must match the email associated with your Glama account. Once published, Glama will automatically detect and verify the file within a few minutes.
Control your server's listing on Glama, including description and metadata
Access analytics and receive server usage reports
Get monitoring and health status updates for your server
Feature your server to boost visibility and reach more users
For users:
Full audit trail – every tool call is logged with inputs and outputs for compliance and debugging
Granular tool control – enable or disable individual tools per connector to limit what your AI agents can do
Centralized credential management – store and rotate API keys and OAuth tokens in one place
Change alerts – get notified when a connector changes its schema, adds or removes tools, or updates tool definitions, so nothing breaks silently
For server owners:
Proven adoption – public usage metrics on your listing show real-world traction and build trust with prospective users
Tool-level analytics – see which tools are being used most, helping you prioritize development and documentation
Direct user feedback – users can report issues and suggest improvements through the listing, giving you a channel you would not have otherwise
The connector status is unhealthy when Glama is unable to successfully connect to the server. This can happen for several reasons:
The server is experiencing an outage
The URL of the server is wrong
Credentials required to access the server are missing or invalid
If you are the owner of this MCP connector and would like to make modifications to the listing, including providing test credentials for accessing the server, please contact support@glama.ai.
Discussions
No comments yet. Be the first to start the discussion!