JobDoneBot
Server Details
84+ free local-first tools: image, PDF, docs, dev utils. Wasm, zero upload, x402 API.
- Status
- Healthy
- Last Tested
- Transport
- Streamable HTTP
- URL
Glama MCP Gateway
Connect through Glama MCP Gateway for full control over tool access and complete visibility into every call.
Full call logging
Every tool call is logged with complete inputs and outputs, so you can debug issues and audit what your agents are doing.
Tool access control
Enable or disable individual tools per connector, so you decide what your agents can and cannot do.
Managed credentials
Glama handles OAuth flows, token storage, and automatic rotation, so credentials never expire on your clients.
Usage analytics
See which tools your agents call, how often, and when, so you can understand usage patterns and catch anomalies.
Tool Definition Quality
Score is being calculated. Check back soon.
Available Tools
170 toolsab-test-title-makerABテストタイトルメーカーAInspect
記事タイトルから10パターン生成。CTRスコア付き (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Discloses 'Browser-based tool' which hints at execution environment, and specifies output quantity (10 patterns). However, with no annotations provided, the description fails to address data privacy (are titles sent to external APIs?), rate limits, or error behaviors, leaving significant behavioral gaps.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Perfectly compact: one front-loaded sentence covering input, output count, and special feature (CTR). Parenthetical '(Browser-based tool)' adds necessary context without verbosity. No wasted words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given zero parameters and no output schema, the description adequately covers the essentials: what goes in (article titles), what comes out (10 variants with scores), and execution context. Missing only output format details (array vs object), but appropriately complete for a simple utility tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Description implies the input is '記事タイトル' (article title), adding crucial semantic information to an otherwise empty input schema. Baseline 4 appropriate for 0-parameter tools, with credit for specifying the expected input resource that the schema omits.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Excellent specificity: states the resource (article titles), action (generates 10 patterns), and distinguishing feature (CTR scores). The 'ABテスト' context in the name combined with these details clearly differentiates it from siblings like article-outline-generator or meta-description-generator.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Provides clear context through specific capabilities (10 variants, CTR scoring) that implicitly signal when to use it (A/B testing headlines for CTR optimization). Lacks explicit 'when not to use' or named alternatives, but the scope is sufficiently specialized to guide selection.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
affiliate-revenue-calcアフィリエイト報酬シミュレーターAInspect
PV×CTR×CVR×単価で月収を即試算。逆算機能付き (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full disclosure burden. It adds crucial behavioral context by noting this is a 'Browser-based tool' (indicating it likely opens a UI rather than returning JSON directly) and mentions the reverse calculation capability. However, it lacks details on persistence, authentication requirements, or whether results can be saved/exported.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is efficiently front-loaded with no waste: one sentence explains the calculation method and formula, a brief clause adds the reverse-calculation feature, and the parenthetical clarifies the delivery mechanism. Every element earns its place.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's low complexity (0 parameters, simple calculator function) and lack of output schema, the description provides sufficient context. It explains what is calculated, how it's calculated, and the browser-based delivery method. A mention of output format (HTML/UI vs JSON) would improve it, but it is adequate for the tool's simplicity.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has zero parameters (empty object), which per guidelines warrants a baseline score of 4. The description adds value by clarifying that PV, CTR, CVR, and 単価 are the conceptual inputs users will provide in the browser interface, compensating for the empty API schema by explaining what data the tool manipulates.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool calculates monthly income (月収) using the specific affiliate marketing formula PV×CTR×CVR×単価 (Page Views × Click-Through Rate × Conversion Rate × Unit Price). It effectively distinguishes this from sibling financial calculators like bond-yield-calc or investment-simulator by specifying the exact e-commerce metrics used.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage context through the formula (use when calculating affiliate revenue from traffic metrics) and mentions the reverse calculation feature (逆算機能付き) for working backwards from revenue targets. While it lacks explicit 'use X instead for Y' comparisons, the specific formula makes the appropriate use case clear relative to other calculators.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
affiliate-revenue-dashboard収益管理ダッシュボードAInspect
案件別×月別の報酬・PV・CVRをグラフ管理。確定申告にも (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description carries full burden. The '(Browser-based tool)' disclosure is essential behavioral information indicating an external UI is opened. However, it omits details on data persistence, authentication requirements, or return format.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Concise Japanese description efficiently structured with core function first, use case second, and implementation detail parenthetically. No wasted words, though slightly more detail on the output behavior would strengthen it.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given zero parameters and no output schema, the description adequately covers the browser-based invocation method but could clarify what data is returned or displayed (e.g., interactive charts vs static data). Sufficient but minimal for a dashboard tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema contains zero parameters. Per evaluation guidelines, 0 params = baseline 4. No parameters require semantic documentation.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Clearly states it manages graphs (グラフ管理) of affiliate rewards, PV, and CVR broken down by project/campaign and month. The 'browser-based' parenthetical and dashboard nature implicitly distinguish it from the sibling calculator tool (affiliate-revenue-calc).
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Mentions tax filing (確定申告) as a specific use case, providing contextual guidance. However, it lacks explicit when-to-use/when-not-to-use rules or references to alternatives like affiliate-revenue-calc for calculation-only needs.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
age-calculator年齢・和暦西暦変換AInspect
生まれ年から年齢・和暦・干支・星座・入学卒業年を一括表示 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full disclosure burden. It adds the crucial behavioral context that this is a 'Browser-based tool', indicating it likely opens a web interface for input rather than accepting function parameters. However, it lacks details on return format, whether results are persisted, or error handling behavior.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely concise single Japanese sentence efficiently conveys input source, all five output types, and interaction model. Every element earns its place with no redundant phrasing or tautology.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's zero-parameter simplicity and lack of output schema, the description provides sufficient context for invocation. It adequately explains what the tool produces and how it accepts input (browser-based). A perfect score would require explicit confirmation of what the tool returns (e.g., whether it returns the calculated data to the agent or merely opens a browser window).
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema has zero parameters, establishing a baseline of 4. The description mentions '生まれ年から' (from birth year), which conceptually explains the required input despite the empty schema. Combined with the 'Browser-based tool' note, this clarifies that birth year input is handled through the interactive UI rather than function arguments, adding necessary semantic context.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description clearly states the tool calculates age, Japanese calendar era (和暦), zodiac (干支), constellation (星座), and school entrance/graduation years from birth year. The specific enumeration of Japanese cultural date systems clearly distinguishes it from sibling tool 'date-calculator'.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The '(Browser-based tool)' annotation implies an interactive usage pattern where input is collected via browser UI rather than parameters, which provides implicit guidance. However, there is no explicit guidance on when to use this versus 'date-calculator' or other calendar tools, and no mention of prerequisites.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
amazon-acos-calculatorACoS/ROAS計算機AInspect
広告費→ACoS/ROAS/損益分岐CPCを即算出 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description carries full burden. Adds valuable context that this is a 'Browser-based tool' (opens interactive UI) and clarifies the input→output flow (ad spend → metrics). However, lacks details on data persistence, rate limiting, or whether calculations are real-time vs cached.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely concise Japanese with arrow notation (→) efficiently mapping inputs to outputs. 'Browser-based tool' parenthetical adds execution context without verbosity. Every element earns its place with zero redundancy.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Adequate for a simple parameterless calculator. The input/output semantics are covered, though it relies on the tool name to establish Amazon context. No output schema exists, but the description enumerates return values (ACoS/ROAS/損益分岐CPC), which suffices for an interactive tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Zero parameters (empty schema) sets baseline to 4 per rubric. Description adds semantic context that the tool expects '広告費' (ad spend) as implicit input and produces ACoS/ROAS metrics, which explains the tool's function despite the empty parameter schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Specific verb '算出' (calculate) with clear resource types: ACoS, ROAS, and break-even CPC from ad spend. The metrics mentioned (ACoS/ROAS) inherently distinguish this from siblings like amazon-fba-calculator (fulfillment fees) and amazon-inventory-calculator (inventory levels).
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The specificity of advertising metrics implies usage for Amazon PPC/advertising analysis, but no explicit when-to-use guidance or named alternatives are provided. For example, it doesn't clarify when to use this vs amazon-keyword-extractor or cvr-improvement-checker.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
amazon-fba-calculatorAmazon FBA料金計算機AInspect
FBA手数料+保管料+配送料→利益を即計算。サイズ区分自動判定 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full behavioral disclosure burden. It mentions '(Browser-based tool)' indicating execution environment, but fails to explain interaction model (e.g., whether it opens an interactive UI, requires manual input, or returns calculated results). It also omits authentication requirements or data persistence details.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely efficient: formula notation (A+B+C→D) conveys calculation scope instantly, size classification feature is highlighted, and browser-based nature is noted parenthetically. Every element earns its place with zero redundancy.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a zero-parameter tool with no output schema, the description adequately explains the calculation purpose. However, the 'Browser-based tool' annotation is cryptic—it should clarify whether this launches an external interface requiring user interaction or operates headlessly. Missing comparison guidance with 10+ sibling calculators.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema contains zero parameters (empty object), which establishes a baseline score of 4 per evaluation rules. The description appropriately makes no parameter claims since none exist, though it could have clarified why no parameters are needed (e.g., 'opens browser for manual input').
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description clearly states the tool calculates FBA fees, storage fees, shipping fees, and profit, using specific domain terminology (サイズ区分/auto size classification). It distinguishes from generic calculators in the sibling list (e.g., base-fee-calculator, mercari-calculator) by explicitly naming Amazon FBA-specific cost components.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
While no explicit 'when-to-use' guidance is provided, the description strongly implies usage context through FBA-specific terminology (size classification, FBA fees) that differentiates it from sibling marketplace calculators. However, it does not explicitly contrast with alternatives like rakuten-fee-calculator or shopify-profit-calculator.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
amazon-inventory-calculatorAmazon在庫回転率計算BInspect
仕入れ数×販売速度→適正在庫・発注タイミングを計算 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
The '(Browser-based tool)' annotation adds crucial behavioral context explaining the empty parameter schema, indicating the tool likely launches an interactive interface rather than accepting calculation inputs via parameters. However, lacking annotations, the description fails to disclose side effects, persistence, or whether it returns a URL vs launching a browser.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely efficient single-line description with formulaic notation (×, →) conveying the calculation logic compactly. No redundant words; every element serves a purpose.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the zero-parameter schema and lack of output schema, the description minimally suffices by explaining the calculation purpose. However, it omits what the tool actually returns (calculation results? a URL? nothing?) and doesn't clarify the browser interaction model beyond the parenthetical hint.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
With zero parameters in the input schema, the baseline score is 4 per evaluation rules. The description mentions the conceptual inputs (purchase quantity, sales velocity) in the calculation formula, which helps explain why the schema is empty (values entered in browser UI), adding semantic value beyond the schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly identifies the calculation being performed (purchase quantity × sales velocity → appropriate inventory/reorder timing) using specific Japanese business terminology. It implicitly distinguishes from sibling tools like amazon-fba-calculator (fees) and amazon-acos-calculator (advertising) by focusing on inventory turnover math, though it could explicitly name relevant siblings.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to select this tool versus similar Amazon calculators (amazon-fba-calculator, amazon-sales-dashboard) or other inventory tools. The description states what it calculates but not under what business conditions it should be invoked.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
amazon-keyword-extractorAmazonキーワード抽出CInspect
商品名からSEOキーワードを分解・サジェスト生成 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description carries full burden. '(Browser-based tool)' hints at UI interaction but fails to disclose what triggers the browser, whether user interaction is required, output format, or side effects. Critical gap given the mutation/implication of scraping behavior.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Single sentence efficiently conveys core function without redundancy. However, given the complexity (no parameters, browser-based implication), the extreme brevity underserves the agent's need for invocation context rather than demonstrating efficient information density.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
With 0 parameters, no annotations, and no output schema, the description should compensate with rich behavioral context. Instead, it leaves critical gaps: how inputs are provided (description implies product names, schema has none), what 'browser-based' execution entails, and what the tool returns.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
While 0 parameters typically baseline at 4, the description creates confusion by stating keywords are extracted 'from product names' (商品名から) despite the empty input schema having no product_name field. This implicit expectation of input without schema support degrades clarity.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Describes specific actions (分解・サジェスト生成/decompose and suggest) on specific resources (SEOキーワード from 商品名), clearly distinguishing from sibling calculators like amazon-fba-calculator. However, 'Browser-based tool' creates slight ambiguity about the execution model.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No explicit guidance on when to use versus alternatives (e.g., when to use this versus amazon-acos-calculator or title-multilingual). No prerequisites or conditions mentioned despite the specialized Amazon SEO domain.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
amazon-sales-dashboardAmazon売上ダッシュボードBInspect
ASIN別売上/利益/在庫をlocalStorageで管理。月別推移グラフ (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Discloses the use of localStorage for data persistence and identifies it as browser-based, which implies client-side operation. However, lacks critical behavioral details given no annotations: storage limits, data lifecycle (what happens on clear), whether it modifies existing data, or if it requires manual data entry versus API sync.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely concise single sentence. Information density is high with no waste. Minor deduction for the parenthetical '(Browser-based tool)' which feels appended rather than integrated, slightly disrupting flow.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Adequately covers the core functionality (ASIN tracking, monthly charts, localStorage) for a parameterless tool, but lacks completeness for a dashboard tool: does not explain how data enters the system (manual input, import, API), export capabilities, or how it differs from the tracking/management siblings in the server.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema has zero parameters. With no parameters to describe, the baseline score of 4 applies. The description appropriately does not reference parameters since none exist.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Clearly states the tool manages Amazon sales, profit, and inventory data by ASIN using localStorage, and identifies it as a browser-based monthly graph dashboard. Distinguishes from sibling calculators by specifying localStorage persistence and ASIN-level granularity.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Provides no explicit guidance on when to use this dashboard versus sibling tools like amazon-inventory-calculator or amazon-fba-calculator. Does not indicate prerequisites for data entry or importing.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
ar-ap-managerAR/AP ManagerBInspect
Accounts receivable/payable tracking with aging analysis. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions 'Browser-based tool,' which hints at a user interface or web access, but does not describe key traits like whether it's read-only or mutative, authentication needs, rate limits, or output format. For a tool with zero annotation coverage, this leaves significant gaps in understanding its behavior.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise and front-loaded: it states the core purpose in one clause and adds a brief technical note in parentheses. Every word earns its place, with no wasted sentences or redundant information, making it easy for an agent to parse quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool has 0 parameters, no output schema, and no annotations, the description is minimally adequate. It covers the basic purpose and a technical aspect ('Browser-based tool'), but lacks details on behavioral traits, usage context, or output expectations. For a tool with no structured data, it provides a foundation but leaves gaps that could hinder effective agent use.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The tool has 0 parameters, and schema description coverage is 100%, so there are no parameters to document. The description does not need to add parameter semantics, and it appropriately avoids unnecessary details. A baseline of 4 is applied as it efficiently handles the lack of parameters without redundancy.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Accounts receivable/payable tracking with aging analysis.' It specifies the verb ('tracking') and resource ('accounts receivable/payable'), and adds a feature ('aging analysis'). However, it does not explicitly differentiate from sibling tools, as none appear to be direct alternatives for AR/AP management, though tools like 'balance-sheet' or 'profit-loss' might be related in financial contexts.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It mentions it is 'Browser-based tool,' which is a technical note but does not offer usage context, prerequisites, or exclusions. Without any when-to-use information, the agent must infer usage from the purpose alone.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
article-outline-generator記事構成ジェネレーターBInspect
KW入力→H2/H3見出し構成を自動提案。テンプレート付き (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden. It adds useful behavioral context by noting it is 'Browser-based' (execution environment) and includes templates (feature behavior), but omits details about data persistence, rate limits, auth requirements, or output format that would be necessary for safe invocation.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely compact single-line description using arrow notation (→) to efficiently convey the input-output flow. Every element (input type, output structure, feature, environment) earns its place with zero redundancy.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given zero schema parameters, no annotations, and no output schema, the description is insufficiently rich. It establishes core functionality but lacks output format details, sibling differentiation, or behavioral safety information needed to compensate for the absent structured metadata.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema is empty (0 parameters), establishing baseline 4. The description semantically clarifies that 'KW' (keyword) input is expected and processed, providing conceptual meaning even though no structured parameters are defined in the schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly defines the transformation (KW input → H2/H3 heading structure proposal) and identifies the resource (article outlines). It specifies distinctive features (templates, browser-based) that differentiate it from generic content generators, though 'KW' abbreviation assumes SEO context without explicitly stating it.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance provided on when to select this tool versus sibling content generators like 'meta-description-generator', 'comparison-table-generator', or 'nda-generator'. No prerequisites, constraints, or alternative selection criteria are mentioned.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
balance-sheetBalance SheetBInspect
Create B/S from assets, liabilities, and equity. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It states 'Create B/S', implying a write or generation operation, but doesn't clarify if this is a read-only simulation, a persistent creation, or what inputs are required (though parameters are zero). It mentions 'Browser-based tool', suggesting it might open a browser interface, but lacks details on permissions, side effects, or output format. For a tool with zero annotation coverage, this is insufficient.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise and front-loaded: 'Create B/S from assets, liabilities, and equity. (Browser-based tool).' Both sentences earn their place—the first defines the core function, and the second adds crucial context about the execution environment. There is no wasted verbiage, making it highly efficient.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (financial statement generation), lack of annotations, no output schema, and zero parameters, the description is minimally complete. It states what the tool does and its browser-based nature, but doesn't explain the output (e.g., a generated document, data structure, or visual), behavioral traits, or how it differs from related financial tools. This is adequate for a simple tool but leaves gaps in understanding its full context.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, meaning no parameters are documented in the schema. The description doesn't add parameter details, which is appropriate since there are none. It implies the tool uses 'assets, liabilities, and equity' as inputs, but these aren't formal parameters. Given zero parameters, a baseline of 4 is applied, as the description doesn't need to compensate for missing schema info.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Create B/S from assets, liabilities, and equity.' It specifies the verb ('Create') and resource ('B/S' meaning Balance Sheet), and distinguishes it from siblings like 'cash-flow-statement' and 'profit-loss' by focusing on balance sheet creation. However, it doesn't explicitly differentiate from 'trial-balance' or 'chart-of-accounts', which might be related financial tools, keeping it from a perfect score.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides minimal usage guidance. It mentions 'Browser-based tool', which hints at the execution environment, but offers no explicit advice on when to use this tool versus alternatives like 'cash-flow-statement' or 'profit-loss'. There are no prerequisites, exclusions, or comparisons to sibling tools, leaving the agent with little context for selection.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
barcode-generatorBarcode GeneratorBInspect
Generate JAN, EAN, Code128 barcodes. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It states the tool generates barcodes and is browser-based, but lacks details on output format (e.g., image type, size), permissions, rate limits, or error handling. For a tool with zero annotation coverage, this is insufficient to inform safe and effective use.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise with two short phrases: 'Generate JAN, EAN, Code128 barcodes. (Browser-based tool).' Every word adds value—specifying barcode types and execution context—with no wasted text, making it front-loaded and efficient.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's simplicity (0 params, no output schema, no annotations), the description is incomplete. It lacks details on output (e.g., image format, download method), behavioral traits like rate limits, and usage context. For a generation tool, this leaves significant gaps in understanding how to invoke and interpret results effectively.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description doesn't add param info, which is appropriate. A baseline of 4 is applied as it meets the requirement for a parameterless tool without redundancy.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Generate JAN, EAN, Code128 barcodes.' It specifies the action (generate) and the resource (barcodes of specific types). However, it doesn't distinguish itself from potential siblings like 'qr-designer' or 'sku-generator' beyond mentioning barcode formats, which is a minor gap.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It mentions 'Browser-based tool,' which hints at execution context but doesn't clarify use cases, prerequisites, or comparisons with sibling tools like 'qr-designer' for different code types. This leaves the agent without explicit usage direction.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
base64-converterBase64 ConverterBInspect
Encode/decode text and files to Base64.
| Name | Required | Description | Default |
|---|---|---|---|
| mode | No | Encode or decode | encode |
| text | Yes | Text to encode or decode | |
| urlSafe | No | Use URL-safe Base64 |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden for behavioral disclosure. While 'encode/decode' implies transformation operations, the description doesn't mention error handling, performance characteristics, limitations (e.g., file size constraints), or what happens with invalid input. For a tool with no annotation coverage, this represents significant gaps in behavioral transparency.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise (5 words) and front-loaded with the core functionality. Every word earns its place with zero waste, making it easy for an agent to quickly understand the tool's purpose.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a relatively simple transformation tool with good schema coverage but no annotations and no output schema, the description provides minimal but adequate context about what the tool does. However, it lacks important details about behavioral characteristics and usage guidelines that would make it more complete for agent invocation.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all three parameters thoroughly. The description mentions 'text and files' which somewhat expands on the 'text' parameter's scope, but doesn't add meaningful semantic context beyond what the schema provides. The baseline score of 3 is appropriate when the schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's function with specific verbs ('encode/decode') and resources ('text and files to Base64'), making the purpose immediately understandable. However, it doesn't differentiate from sibling tools, which appear to be unrelated utilities rather than competing Base64 converters, so sibling differentiation isn't applicable here.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives, prerequisites, or contextual considerations. It simply states what the tool does without any usage context, leaving the agent to infer appropriate scenarios.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
base-fee-calculatorBASE手数料計算機AInspect
サービス利用料3%+決済手数料→利益計算。スタンダード/グロース比較 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden. It successfully discloses the browser-based nature of the tool (important for agent expectation management) and specifies the calculation logic (3% service fee + payment fees). It stops short of explaining side effects, persistence, or exact return behavior, but covers the essential behavioral context.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely efficient: 'サービス利用料3%+決済手数料→利益計算' defines the calculation logic, 'スタンダード/グロース比較' identifies the comparison feature, and '(Browser-based tool)' indicates the modality. No wasted words, front-loaded with core functionality.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a zero-parameter calculator tool without output schema, the description is sufficiently complete. It identifies the platform (BASE), fee structure, and browser-based delivery method. It could optionally clarify that no input parameters are required, though this is evident from the schema.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema contains zero parameters. Per rubric, this establishes a baseline of 4. The description appropriately makes no mention of parameters (which would be confusing given the empty schema), maintaining consistency with the schema definition.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description clearly states the tool calculates BASE platform fees (3% service fee + payment processing fee) for profit calculation and compares Standard vs Growth plans. The specific platform name (BASE) and fee structure distinguish it from sibling calculators like rakuten-fee-calculator and amazon-fba-calculator.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
While the platform-specific name and 'Standard/Growth comparison' implicitly suggest use for BASE fee analysis, there is no explicit guidance on when to select this over the numerous sibling fee calculators (e.g., mercari-calculator, shopify-profit-calculator) or prerequisites for use.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
bg-removerBackground RemoverBInspect
One-click AI background removal for people, products, and objects. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden but only adds 'Browser-based tool' as behavioral context. It lacks crucial details: whether it's read-only or destructive, authentication needs, rate limits, output format, or processing time. For a zero-parameter AI tool, this leaves significant behavioral gaps.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely concise with two brief phrases that each add value: the first states the core function and scope, the second provides implementation context. No wasted words, perfectly front-loaded with the primary purpose.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a zero-parameter tool with no output schema and no annotations, the description is incomplete. It doesn't explain what happens after clicking (e.g., file upload interface, output format, success/failure behavior). The 'Browser-based tool' hint helps but doesn't compensate for missing operational details.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The tool has zero parameters with 100% schema description coverage, so the schema already fully documents the input interface. The description appropriately doesn't add parameter details, maintaining focus on the tool's purpose rather than redundant information.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's function as 'One-click AI background removal' with specific targets (people, products, objects), which goes beyond just restating the name/title. However, it doesn't explicitly differentiate from sibling tools like 'pro-matting' or 'image-crop' that might have overlapping functionality.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides minimal guidance with 'Browser-based tool' hinting at context, but offers no explicit when-to-use advice, prerequisites, or alternatives. It doesn't help an agent decide between this and similar sibling tools like 'pro-matting' or 'image-crop'.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
bond-yield-calcBond Yield CalcCInspect
Yield to maturity, duration, and convexity calculator.
| Name | Required | Description | Default |
|---|---|---|---|
| calcMode | No | Calculation mode | price-to-yield |
| faceValue | No | Face value (per 100 yen) | |
| frequency | No | Coupon payment frequency | semi-annual |
| couponRate | No | Coupon rate (%) | |
| targetYield | No | Target yield (%) for yield-to-price mode | |
| currentPrice | No | Current market price (per 100 yen) | |
| yearsToMaturity | No | Years to maturity |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It states the tool is a 'calculator,' implying a read-only, computational function without side effects, but does not explicitly confirm this or detail any behavioral traits like error handling, performance limits, or output format. For a tool with no annotations, this is a significant gap in transparency.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise with a single sentence: 'Yield to maturity, duration, and convexity calculator.' It is front-loaded and wastes no words, making it easy to parse quickly. Every part of the sentence directly contributes to understanding the tool's purpose.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (7 parameters, financial calculations) and lack of annotations and output schema, the description is insufficient. It does not explain the relationship between inputs and outputs, error conditions, or practical usage examples. For a calculator with multiple modes and parameters, more context is needed to guide effective use.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The description does not add any parameter-specific information beyond what the input schema provides. Since schema description coverage is 100%, the baseline score is 3. The description's mention of 'yield to maturity, duration, and convexity' hints at outputs but does not clarify input semantics, such as how parameters like 'calcMode' affect calculations, so it adds no extra value over the schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose as a calculator for 'yield to maturity, duration, and convexity,' which are specific financial metrics for bonds. It uses the verb 'calculator' with the resource 'bond yield' implied in the name, making the function unambiguous. However, it does not explicitly differentiate from sibling tools like 'npv-irr-calc' or 'investment-simulator,' which might also involve financial calculations, so it misses full sibling distinction.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It does not mention any prerequisites, context, or exclusions, such as when to choose this over other financial calculators in the sibling list. The lack of usage context leaves the agent without direction on appropriate scenarios for invocation.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
career-history-generatorCareer HistoryBInspect
Professional CV in 3 format styles. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It mentions the tool is browser-based, which adds some context about the execution environment, but fails to describe output format, whether it requires user input, or any limitations (e.g., data persistence, privacy). This leaves significant gaps for a tool that likely generates documents.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise—two short phrases that are front-loaded with the core functionality. Every word earns its place, with no redundant or vague language.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool has 0 parameters and no output schema, the description is minimally adequate. It covers the basic purpose and execution context, but as a document-generation tool with no annotations, it should ideally specify output types or usage constraints to be fully complete.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description adds value by specifying '3 format styles' and 'browser-based', which provides context beyond the empty schema, though it doesn't detail what those formats are.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool generates a professional CV in 3 format styles, specifying both the resource (CV) and the action (generate). It distinguishes from siblings like 'resume-generator' by mentioning format styles, though it could be more explicit about the difference.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives like 'resume-generator' or other career-related tools. The description mentions it's browser-based, which is a minor contextual hint, but lacks explicit usage scenarios or exclusions.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
cash-flow-statementCash Flow StatementCInspect
Indirect method cash flow statement generation. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It states the tool is 'browser-based', which hints at execution environment but doesn't disclose key behavioral traits like whether it requires user input, how it generates the statement, if it's read-only or creates outputs, or any limitations. For a tool with zero annotation coverage, this leaves significant gaps in understanding its operation.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is very brief—two short phrases—with no wasted words. It's front-loaded with the core purpose. However, it could be slightly more informative without losing conciseness, such as clarifying what 'indirect method' entails or the output format.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no annotations, no output schema, and a tool that likely produces a financial statement (a non-trivial output), the description is incomplete. It doesn't explain what the tool returns, how it operates, or any dependencies. For a generation tool with zero structured context, this leaves too many unanswered questions for effective use.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description doesn't add param info, which is appropriate here. Baseline is 4 for 0 parameters, as the schema fully covers the absence of inputs.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description states it's for 'cash flow statement generation' using the 'indirect method', which clarifies the verb (generate) and resource (cash flow statement) with a specific method. However, it doesn't distinguish from sibling tools like 'balance-sheet' or 'profit-loss' beyond the different financial statement type, leaving some ambiguity about its unique scope compared to other financial tools.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives. The description mentions it's 'browser-based', but this doesn't help differentiate it from other tools or indicate specific contexts or prerequisites for use. There's no mention of when-not-to-use or comparisons to similar financial tools.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
certificate-makerCertificate MakerBInspect
Award certificates with templates and name mail-merge. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions 'Browser-based tool', indicating it likely opens a browser interface, but doesn't cover critical aspects like whether it creates permanent outputs, requires authentication, has rate limits, or what happens after execution. For a tool with potential file generation, this is a significant gap.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise—two brief phrases that directly state the tool's function and execution context. Every word earns its place, with no redundancy or fluff, making it easy to parse quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (likely involves document generation) and lack of annotations or output schema, the description is minimally adequate. It covers the core purpose and execution method but omits details on outputs, errors, or behavioral traits. For a zero-parameter tool, it meets basic needs but leaves gaps for informed use.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description adds value by hinting at functionality ('templates and name mail-merge'), which suggests internal options or UI elements, though it doesn't detail them. Baseline is 4 for zero parameters, as the description provides some context beyond the empty schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose as 'Award certificates with templates and name mail-merge', specifying the action (award), resource (certificates), and key features (templates, mail-merge). It distinguishes itself from siblings like 'resume-generator' or 'contract-generator' by focusing on certificates, though it doesn't explicitly differentiate from similar document-generation tools.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives. The description mentions it's 'Browser-based', which hints at a UI context, but doesn't specify prerequisites, ideal scenarios, or exclusions. Without this, users must infer usage from the purpose alone.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chapter-list-generatorチャプターリスト生成BInspect
動画の区切りを入力→YouTube用チャプター形式で出力 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Without annotations, the description carries the full disclosure burden. It adds valuable context by labeling itself 'Browser-based tool' (indicating client-side execution/UI interaction), but fails to specify output format details, input size constraints, or whether the output is copied to clipboard vs returned as text.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely efficient: input→output format logic with a parenthetical implementation detail. Every element earns its place; no redundancy. The arrow notation effectively conveys the transformation in minimal space.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a simple formatting utility with no output schema, the description covers the core purpose but remains incomplete. It should specify the exact YouTube chapter syntax (e.g., '00:00 Title') and clarify how the browser-based interaction works (opens input dialog? requires manual paste?).
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
With zero parameters, baseline is 4 per scoring rules. The description adds necessary semantic context by describing what input the tool expects ('動画の区切り') even though it is not provided via parameters, clarifying the tool's interactive nature.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the transformation workflow (video segments → YouTube chapter format) using specific terminology. However, it could better distinguish from `timestamp-generator` and `video-duration-calculator` by explicitly mentioning it formats timestamps with titles into the specific 00:00 syntax used by YouTube.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
While the arrow notation implies a workflow, there is no explicit guidance on when to use this versus siblings like `youtube-description-generator` or `timestamp-generator`. It does not state prerequisites (e.g., having timestamp data ready) or that it is an interactive browser tool versus API-style.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chart-of-accountsChart of AccountsCInspect
100+ standard accounts with search and custom additions. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It mentions 'search and custom additions', hinting at read and write capabilities, but doesn't clarify permissions, data persistence, or side effects. The 'Browser-based tool' note adds implementation detail but not operational behavior, leaving significant gaps for a tool that likely modifies financial data.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is brief and front-loaded with the core functionality ('100+ standard accounts with search and custom additions'), though the parenthetical note adds minor clutter. It avoids redundancy and stays focused, but could be slightly more polished by integrating the browser note more seamlessly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity of financial account management, no annotations, and no output schema, the description is insufficient. It lacks details on what the tool returns, how 'search' and 'custom additions' work, error conditions, or integration with sibling tools, making it incomplete for safe and effective use.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The tool has 0 parameters with 100% schema description coverage, so no parameter documentation is needed. The description doesn't discuss parameters, which is appropriate here, earning a baseline score of 4 for not introducing unnecessary information.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description '100+ standard accounts with search and custom additions' vaguely indicates a tool for managing a chart of accounts, but lacks a specific verb (e.g., 'view', 'manage', 'edit') and doesn't clearly distinguish it from sibling tools like 'balance-sheet' or 'trial-balance'. The parenthetical '(Browser-based tool)' adds technical context but not functional clarity.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives like 'balance-sheet' or 'journal-entry'. The description implies it handles accounts, but doesn't specify use cases, prerequisites, or exclusions, leaving the agent to infer usage from the tool name alone.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
check-subscriptionCheck Subscription StatusCInspect
Check the status of a JobDoneBot checkout session.
| Name | Required | Description | Default |
|---|---|---|---|
| session_id | Yes | Stripe checkout session ID (cs_xxx) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It states the action ('Check') but doesn't disclose behavioral traits like authentication needs, rate limits, error conditions, or what the output might contain. For a tool that likely interacts with external services (Stripe), this is a significant gap in transparency.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence with zero waste. It's front-loaded with the core purpose and avoids unnecessary elaboration, making it easy to parse quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a tool with no annotations and no output schema, the description is incomplete. It doesn't explain what 'status' entails (e.g., active, canceled, pending), the format of the response, or any dependencies on external systems like Stripe. This leaves critical gaps for an agent to use the tool effectively.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The schema description coverage is 100%, with the single parameter 'session_id' well-documented in the schema as 'Stripe checkout session ID (cs_xxx)'. The description adds no additional parameter semantics beyond what the schema provides, so the baseline score of 3 is appropriate.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the verb ('Check') and resource ('status of a JobDoneBot checkout session'), making the purpose specific and understandable. However, it doesn't distinguish this tool from potential siblings like 'subscribe-pro' or other subscription-related tools, which could cause confusion in a crowded toolset.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives, prerequisites, or context. Given the many sibling tools (e.g., 'subscribe-pro'), this lack of differentiation leaves the agent guessing about appropriate use cases.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
classroom-timerClassroom TimerBInspect
Full-screen countdown timer with chime sounds. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden for behavioral disclosure. It mentions the browser-based nature and chime sounds, but doesn't cover important aspects like whether it requires user interaction, how the timer is controlled, what happens when it completes, or any limitations (e.g., maximum duration, sound options). For a tool with zero annotation coverage, this leaves significant behavioral gaps.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise at just two phrases, with zero wasted words. It's front-loaded with the core functionality ('full-screen countdown timer with chime sounds') followed by implementation context ('browser-based tool'). Every element earns its place.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a zero-parameter tool with no output schema, the description provides basic purpose and implementation context. However, without annotations and given the behavioral gaps mentioned, it's only minimally adequate. The agent knows what the tool does at a high level but lacks important operational details.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The tool has 0 parameters with 100% schema description coverage, so the schema already fully documents the parameter situation. The description appropriately doesn't discuss parameters, maintaining focus on the tool's purpose and behavior. Baseline for 0 parameters is 4.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's function as a 'full-screen countdown timer with chime sounds' and specifies it's 'browser-based', which provides specific verb+resource+environment context. However, it doesn't differentiate from the sibling 'countdown-timer' tool, which appears to serve a similar purpose.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided about when to use this tool versus alternatives. The description doesn't mention any specific use cases, prerequisites, or comparisons with the sibling 'countdown-timer' tool, leaving the agent without contextual usage direction.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
code-formatterCode FormatterCInspect
Format source code using Prettier. Supports JS, TS, CSS, HTML, JSON, YAML, Markdown, GraphQL.
| Name | Required | Description | Default |
|---|---|---|---|
| code | Yes | Source code to format | |
| semi | No | Add semicolons (JS/TS) | |
| useTabs | No | Use tabs instead of spaces | |
| language | No | Programming language | typescript |
| tabWidth | No | Tab width in spaces | |
| printWidth | No | Line width before wrapping | |
| singleQuote | No | Use single quotes (JS/TS) | |
| trailingComma | No | Trailing comma style | es5 |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full behavioral burden but fails to disclose error handling (what happens on invalid syntax), whether output is returned as string or file, or the read-only/transformative nature of the operation beyond the implied 'format' verb.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences with zero waste. Front-loaded with action ('Format source code') and immediately qualifies scope (engine + languages). Every word earns its place.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Lacks output schema and annotations, yet description fails to explain return values, error states, or success behavior. For an 8-parameter transformation tool, the absence of output documentation creates a significant gap in agent understanding.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%, establishing baseline 3. Description adds 'Prettier' context which helps frame the style options (semi, trailingComma) and lists supported languages matching the enum, but adds no specific syntax guidance or parameter relationships beyond schema definitions.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
States specific verb (Format) + resource (source code) + engine (Prettier) and lists supported languages. Implicitly distinguishes from sibling 'sql-formatter' by omitting SQL and from 'format-converter' by specifying Prettier/source code, though lacks explicit comparative guidance.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Provides no guidance on when to select this tool versus siblings like 'sql-formatter' (SQL formatting) or 'format-converter' (file conversion), nor mentions prerequisites like valid syntax requirements.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
color-converterカラー変換CInspect
HEX・RGB・HSL・CMYKを相互変換。カラーピッカー付き
| Name | Required | Description | Default |
|---|---|---|---|
| input | Yes | Color value (e.g. "#ff0000", "255,0,0", "0,100,50") | |
| fromFormat | No | Input color format | hex |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It mentions 'カラーピッカー付き' (with color picker), hinting at additional functionality, but doesn't disclose behavioral traits like whether it's read-only, requires authentication, has rate limits, or what the output looks like. For a conversion tool with zero annotation coverage, this is inadequate.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely concise with two brief phrases: 'HEX・RGB・HSL・CMYKを相互変換' and 'カラーピッカー付き'. Every word earns its place, and the purpose is front-loaded. No wasted sentences.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no annotations and no output schema, the description is incomplete. It doesn't explain what the tool returns (e.g., converted values, error handling), behavioral constraints, or usage context. For a tool with two parameters and conversion logic, more detail is needed to guide an AI agent effectively.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema fully documents both parameters. The description doesn't add any parameter-specific information beyond what's in the schema. It implies conversion between formats but doesn't explain parameter interactions or provide examples beyond the schema's examples. Baseline 3 is appropriate when schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's function: 'HEX・RGB・HSL・CMYKを相互変換' (mutual conversion between HEX, RGB, HSL, CMYK formats). It specifies the verb (convert) and resources (color formats). However, it doesn't explicitly distinguish from sibling tools like 'color-palette' or 'image-color-picker' beyond mentioning 'カラーピッカー付き' (with color picker).
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives. The description doesn't mention sibling tools or contextual usage scenarios. It lacks prerequisites, exclusions, or comparisons to similar tools in the list.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
color-paletteColor PaletteBInspect
Generate harmonious color palettes for design projects. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It mentions 'browser-based tool,' which hints at client-side execution but doesn't clarify key traits like whether it's deterministic, if it requires internet access, what the output format is (e.g., HEX codes, RGB values), or any limitations (e.g., number of colors per palette). This leaves significant gaps for an AI agent to understand how to invoke it effectively.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise—two brief phrases that directly state the tool's function and context. Every word earns its place: 'Generate harmonious color palettes' defines the action, 'for design projects' provides usage context, and '(Browser-based tool)' adds implementation detail. No wasted sentences or redundancy.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's simplicity (0 parameters, no annotations, no output schema), the description is minimally adequate. It covers the basic purpose and hints at execution environment but lacks details on output format, behavioral constraints, or integration guidelines. For a tool with no structured data to rely on, it should do more to be fully helpful, but it meets the bare minimum for a zero-parameter tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description doesn't add parameter details, which is appropriate here. A baseline of 4 is given since the schema fully handles parameters, and the description doesn't need to compensate—it focuses on the tool's purpose instead.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with a specific verb ('Generate') and resource ('harmonious color palettes'), and provides context ('for design projects'). It distinguishes from siblings like 'color-converter' or 'image-color-picker' by focusing on palette generation rather than conversion or extraction. However, it doesn't explicitly differentiate from all possible palette-related tools, keeping it at a 4 rather than a 5.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides minimal usage guidance—only that it's for design projects and browser-based. It doesn't specify when to use this tool versus alternatives (e.g., 'color-converter' for single colors, 'image-color-picker' for extracting from images) or any prerequisites. No explicit when/when-not instructions are given, making it inadequate for informed tool selection.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
comparison-table-generator比較表ジェネレーターAInspect
商品比較表HTMLを30秒で生成。コピペでWPに貼れる (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden and discloses key behavioral traits: performance ('30秒で' - 30 seconds) and execution environment ('Browser-based tool'). However, it omits details about data persistence, offline capability, or output limitations that would be valuable for a zero-parameter tool.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely efficient two-sentence structure with zero waste. The first sentence covers function and speed; the second covers integration workflow and technical context. Information is front-loaded with the core value proposition (HTML generation).
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's simplicity (zero parameters) and lack of output schema, the description adequately covers the essential context: output format (HTML), usage workflow (copy-paste), and performance expectations. It sufficiently enables an agent to recommend this tool for WordPress content creation scenarios.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has zero parameters, which per the rubric establishes a baseline of 4. The description appropriately makes no parameter claims since the tool appears to work without inputs or uses an internal wizard, matching the empty schema structure.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description uses specific action '生成' (generate) with clear resource '商品比較表HTML' (product comparison table HTML). It effectively distinguishes from sibling generators like 'article-outline-generator' or 'contract-generator' by specifying the exact output type (comparison tables) and target platform (WP/WordPress).
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The phrase 'コピペでWPに貼れる' implies the intended use case (creating content for WordPress paste), providing implicit context. However, it lacks explicit guidance on when NOT to use this versus other table tools or prerequisites for use.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
contract-generatorContract GeneratorCInspect
Generate business contracts from templates. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It mentions 'Browser-based tool', hinting at UI interaction, but fails to disclose critical behaviors like whether it creates editable documents, requires authentication, has rate limits, or handles errors. This leaves significant gaps for a generation tool.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise with two short phrases, front-loading the core purpose ('Generate business contracts from templates') and adding a brief note on implementation ('Browser-based tool'). Every word serves a purpose without redundancy.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a tool with no annotations, no output schema, and 0 parameters, the description is incomplete. It lacks details on output format (e.g., PDF, editable document), error handling, template sources, or any behavioral traits needed for effective use in a business context.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
With 0 parameters and 100% schema coverage, the baseline is 4. The description doesn't need to add parameter details, but it implies templates are involved without specifying how they're selected or customized, which is a minor gap given the zero-param context.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description 'Generate business contracts from templates' clearly states the tool's action (generate) and resource (business contracts), but it's vague about scope and templates. It doesn't distinguish from siblings like 'nda-generator' or 'terms-generator', which might serve similar legal document purposes.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance on when to use this tool versus alternatives like 'nda-generator' or 'terms-generator' is provided. The description lacks context about prerequisites, such as needing templates or specific business contexts, offering no usage differentiation.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
countdown-timerカウントダウンAInspect
クリスマス・正月・GWまであと何日?リアルタイムカウントダウン (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It adds 'リアルタイム' (real-time) and 'Browser-based tool' which hint at the operational nature, but fails to specify whether it returns structured data, opens a browser window, or requires user interaction, leaving significant ambiguity for agent invocation.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely compact with zero redundancy: a question establishing the use case followed by a two-word functional label and parenthetical technical note. Every component earns its place without filler.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the absence of an output schema and annotations, the description should ideally disclose what the tool returns (e.g., days remaining, a visual timer, or a browser URL). While it mentions the browser-based nature, it leaves the actual output format and side effects unspecified, creating a gap for agent decision-making.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema contains zero parameters, establishing a baseline score of 4. The description appropriately does not invent parameter documentation where none exists, matching the schema structure.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly identifies the tool's function (countdown to specific dates/holidays) using concrete examples (Christmas, New Year, Golden Week). It distinguishes from siblings like classroom-timer and timezone-converter by focusing on holiday/event countdowns rather than general timing or conversion tasks.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage scenarios through the listed holiday examples (Christmas, New Year, GW), giving users context for when to invoke this tool. However, it lacks explicit guidance on when not to use it or alternatives for custom date countdowns.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
cron-generatorCron GeneratorCInspect
Build and parse cron expressions visually.
| Name | Required | Description | Default |
|---|---|---|---|
| mode | No | Parse existing expression or generate description | parse |
| expression | No | Cron expression (5-field format: minute hour day month weekday) | * * * * * |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden but reveals little. 'Visually' hints at a user-friendly interface, but doesn't disclose behavioral traits like whether it's read-only vs. mutating, authentication needs, rate limits, error handling, or what 'visually' entails. For a tool with zero annotation coverage, this is insufficient.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely concise with a single, front-loaded sentence that wastes no words. Every word ('Build', 'parse', 'cron expressions', 'visually') contributes directly to understanding the tool's core functionality.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no annotations, no output schema, and a tool that likely produces visual or descriptive output, the description is incomplete. It doesn't explain what 'visually' means (e.g., graphical interface, human-readable description), what the output looks like, or any constraints. For a tool with dual modes and visual output, this lacks necessary context.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents both parameters ('mode' with enum values, 'expression' with format). The description adds no parameter-specific details beyond what's in the schema, such as examples of valid expressions or visual output format. Baseline 3 is appropriate when schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description 'Build and parse cron expressions visually' clearly states the tool's dual functionality with specific verbs ('build', 'parse') and resource ('cron expressions'). It distinguishes from most siblings by focusing on cron expressions, though it doesn't explicitly differentiate from potential similar tools like 'timestamp-converter' or 'date-calculator'.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives. The description doesn't mention prerequisites, typical use cases, or when to choose 'parse' vs 'generate' mode. It's a generic statement that offers no contextual decision-making help.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
csv-excel-converterCSV-Excel ConverterBInspect
Convert between CSV and Excel (.xlsx) in the browser.
| Name | Required | Description | Default |
|---|---|---|---|
| csv | No | CSV text content (required for csv-to-excel) | |
| excel | No | Base64-encoded Excel file (required for excel-to-csv) | |
| delimiter | No | CSV delimiter | , |
| direction | Yes | Conversion direction | |
| sheetIndex | No | Sheet index for excel-to-csv (0-based) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions the conversion happens 'in the browser', which adds some context about the execution environment. However, it lacks details on permissions, rate limits, error handling, or what the output looks like (e.g., format of returned data). For a tool with no annotations, this is insufficient.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence that front-loads the core functionality. There's no wasted text, and it's appropriately sized for the tool's complexity. Every word earns its place by clearly stating the conversion purpose and context.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (5 parameters, no output schema, no annotations), the description is minimally adequate. It covers the basic purpose but lacks details on behavioral traits, output format, or usage guidelines. With no output schema, it should ideally hint at return values, but the schema's high coverage mitigates some gaps, making it just viable.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, meaning all parameters are documented in the schema. The description doesn't add any parameter-specific details beyond what's in the schema (e.g., it doesn't explain the 'direction' parameter's implications or provide examples). With high schema coverage, the baseline is 3, as the description doesn't compensate with extra semantic value.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: converting between CSV and Excel formats in the browser. It specifies the verb 'convert' and resources 'CSV and Excel (.xlsx)', making the function unambiguous. However, it doesn't differentiate from sibling tools like 'json-to-csv' or 'format-converter', which might handle similar conversions, so it doesn't reach a perfect score.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It doesn't mention sibling tools like 'json-to-csv' or 'format-converter', nor does it specify prerequisites or exclusions. The usage is implied by the name and description, but no explicit context is given.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
csv-queryCSV QueryCInspect
Run SQL-like queries on CSV data.
| Name | Required | Description | Default |
|---|---|---|---|
| csv | Yes | CSV data | |
| query | Yes | SQL-like query (e.g. "SELECT name, age WHERE age > 30 ORDER BY name LIMIT 10") |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden but offers minimal behavioral insight. It mentions 'SQL-like queries' but doesn't disclose critical traits like whether this is a read-only operation, error handling, performance constraints, or output format. For a tool that processes data, this leaves significant gaps in understanding its behavior.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise—a single sentence with zero wasted words. It's front-loaded with the core functionality ('Run SQL-like queries on CSV data'), making it easy to parse quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (querying data), lack of annotations, and no output schema, the description is insufficient. It doesn't explain what the tool returns (e.g., results, errors), behavioral constraints, or usage context, leaving the agent with inadequate information to use it effectively.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents both parameters thoroughly. The description adds no additional parameter semantics beyond what's in the schema (e.g., CSV format details or query syntax examples), meeting the baseline for high schema coverage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with a specific verb ('Run') and resource ('CSV data'), making it immediately understandable. However, it doesn't distinguish this tool from potential siblings like 'csv-excel-converter' or 'json-to-csv' that also handle CSV data but with different operations.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided about when to use this tool versus alternatives. The description doesn't mention prerequisites (e.g., CSV format requirements), limitations (e.g., query complexity), or related tools (e.g., 'csv-excel-converter' for format conversion instead of querying).
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
cvr-improvement-checkerCVR改善チェッカーBInspect
記事のCTA配置・リンク密度・読了率を分析。成約率UPの改善点を提示 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It identifies the browser-based nature of the tool, which explains the empty parameter schema, but fails to disclose what data sources it accesses, privacy implications of analyzing content, or the format of improvement suggestions returned.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The two-sentence Japanese description is tightly constructed with no wasted words. The first sentence covers analysis targets and the second covers output deliverables, with the parenthetical browser notation earning its place by explaining the parameter-less schema.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a zero-parameter tool without output schema, the description adequately covers the core functionality. However, it lacks critical contextual details such as how the article content is provided (context, clipboard, URL?), expected processing time, or the nature of the browser interaction, leaving gaps in the agent's ability to predict execution behavior.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema contains zero parameters, which per evaluation rules establishes a baseline score of 4. The description implicitly explains why the schema is empty (browser-based tool handles inputs separately), providing adequate semantic context for the parameter structure.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description uses specific verbs (分析, 提示) and resources (CTA配置, リンク密度, 読了率) to clearly define the tool's focus on conversion rate optimization for articles. It distinguishes from siblings like article-outline-generator (creation) and sales-writing-analyzer (general analysis) through specific metrics mentioned.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
While the '(Browser-based tool)' notation hints at the execution mechanism, the description provides no explicit guidance on when to use this tool versus similar alternatives like sales-writing-analyzer or rewrite-assistant, nor does it mention prerequisites such as requiring article content or URLs.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
data-fakerData FakerBInspect
Generate realistic fake data (names, emails, addresses) in JSON/CSV.
| Name | Required | Description | Default |
|---|---|---|---|
| seed | No | Random seed for reproducible output | |
| count | No | Number of records to generate | |
| fields | Yes | Field definitions | |
| locale | No | Data locale | ja |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions output formats (JSON/CSV) but does not cover critical aspects like rate limits, authentication requirements, error handling, or whether the operation is idempotent. For a data generation tool with no annotation coverage, this leaves significant gaps in understanding its behavior.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise and front-loaded, consisting of a single sentence that directly states the tool's purpose and output formats. There is no wasted language, and every word contributes to understanding the tool's function.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (4 parameters, no output schema, and no annotations), the description is minimally adequate. It covers the basic purpose and output formats but lacks details on behavioral traits, usage guidelines, and parameter nuances. The high schema description coverage helps, but the overall context remains incomplete for effective tool selection and invocation.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The description does not add any parameter-specific information beyond what is already in the input schema, which has 100% description coverage. It mentions the types of data generated (e.g., names, emails, addresses), which loosely relates to the 'fields' parameter but does not provide additional semantics or usage examples. The baseline score of 3 is appropriate given the high schema coverage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Generate realistic fake data (names, emails, addresses) in JSON/CSV.' It specifies the verb ('Generate'), resource ('realistic fake data'), and output formats. However, it does not explicitly differentiate from sibling tools, as none appear to be direct alternatives for fake data generation.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It lacks context about prerequisites, typical use cases, or comparisons with other tools. Without such information, the agent must infer usage based solely on the tool's name and description.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
date-calculatorDate CalculatorBInspect
Calculate date differences, add/subtract dates, count business days.
| Name | Required | Description | Default |
|---|---|---|---|
| mode | No | Calculation mode | diff |
| unit | No | Unit for add mode | |
| date1 | Yes | First date (YYYY-MM-DD) | |
| date2 | No | Second date (YYYY-MM-DD) for diff mode | |
| amount | No | Amount to add/subtract (for add mode) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It lists operations but doesn't specify input formats (though the schema covers this), error handling, timezone handling, or output structure. For a tool with multiple modes and no annotations, this leaves significant behavioral gaps.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise—a single sentence with three comma-separated clauses—and front-loaded with all key operations. Every word earns its place, with no redundant or vague phrasing.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (5 parameters, 3 modes), 100% schema coverage, and no output schema, the description is minimally adequate. It lists operations but lacks details on output format, error cases, or behavioral nuances. Without annotations or output schema, more context would be helpful for effective use.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema fully documents all parameters. The description mentions 'date differences', 'add/subtract dates', and 'count business days', which loosely map to the 'mode' parameter values ('diff', 'add', 'business-days'), but adds no additional semantic context beyond what the schema provides. This meets the baseline for high schema coverage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with specific verbs ('calculate', 'add/subtract', 'count') and resources ('date differences', 'dates', 'business days'). It distinguishes itself from siblings like 'age-calculator' and 'timestamp-converter' by focusing on date arithmetic rather than age calculation or timestamp conversion, though it doesn't explicitly differentiate from them.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It doesn't mention prerequisites, exclusions, or compare it to sibling tools like 'age-calculator' or 'timestamp-converter', leaving the agent to infer usage context solely from the tool name and description.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
delivery-note-generatorDelivery NoteBInspect
Generate delivery notes with receipt confirmation field. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It mentions 'Browser-based tool,' indicating it runs in a browser context, which is useful. However, it lacks details on permissions, side effects (e.g., whether it saves or downloads files), rate limits, or output format. For a generation tool with zero annotation coverage, this leaves significant behavioral gaps.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is brief and front-loaded with the main purpose, but the second part 'Browser-based tool' feels tacked on without clear necessity. It's concise but could be more structured to integrate the context seamlessly, earning a mid-range score for minimal waste but room for improvement.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given zero parameters and no output schema, the description is moderately complete. It states what the tool does and adds browser context, but as a generation tool with no annotations, it should ideally mention output behavior (e.g., file download, preview). It's adequate but has gaps in behavioral transparency.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The tool has 0 parameters with 100% schema description coverage, so the schema fully documents inputs. The description adds no parameter information, which is acceptable here. Baseline is 4 for zero parameters, as there's nothing to compensate for, and the schema suffices.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Generate delivery notes with receipt confirmation field.' It specifies the verb ('Generate'), resource ('delivery notes'), and a key feature ('receipt confirmation field'). However, it doesn't differentiate from sibling tools like 'invoice-generator' or 'receipt-generator' that might overlap in document generation contexts.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It mentions 'Browser-based tool,' which hints at the execution environment but doesn't clarify use cases, prerequisites, or comparisons to siblings like 'invoice-generator' or 'purchase-order' that might serve similar business documentation needs.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
depreciation-calcDepreciation CalcCInspect
Straight-line and declining-balance depreciation schedules.
| Name | Required | Description | Default |
|---|---|---|---|
| method | No | Depreciation method | straight-line |
| usefulLife | Yes | Useful life in years | |
| acquisitionCost | Yes | Acquisition cost in JPY | |
| assetCategoryId | No | Asset category ID (optional) | |
| serviceStartDate | Yes | Service start date (YYYY-MM format) | |
| fiscalYearEndMonth | No | Fiscal year end month (1-12) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions the calculation methods but does not describe output format (e.g., schedule structure, time periods), error handling, or computational limits. For a tool with no annotations, this leaves significant gaps in understanding how it behaves.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise—a single phrase that directly states the tool's core functionality with zero redundant words. It is front-loaded and efficiently communicates the essential purpose without unnecessary elaboration.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (financial calculations with 6 parameters) and lack of annotations or output schema, the description is insufficient. It does not explain the output format, error conditions, or provide examples, making it hard for an agent to use correctly without trial and error.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema fully documents all parameters. The description adds no additional parameter semantics beyond what the schema provides, such as explaining relationships between parameters (e.g., how method affects calculations) or typical values. This meets the baseline for high schema coverage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: generating depreciation schedules using straight-line and declining-balance methods. It specifies the verb ('depreciation schedules') and resource (implied financial calculations), but does not differentiate from sibling tools like 'npv-irr-calc' or 'bond-yield-calc' which are also financial calculators, leaving some ambiguity about when to choose this specific tool.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It does not mention prerequisites, typical use cases, or compare it to sibling financial calculation tools, leaving the agent with no context for tool selection beyond the name and description.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
diff-checkerDiff CheckerBInspect
Side-by-side text and code diff comparison.
| Name | Required | Description | Default |
|---|---|---|---|
| left | Yes | Left/original text to compare | |
| right | Yes | Right/modified text to compare | |
| format | No | Output format (unified=git-style diff, json=structured, stats=summary only) | unified |
| leftName | No | Label for left/original text | Original |
| rightName | No | Label for right/modified text | Modified |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions 'side-by-side' comparison, implying a visual or structured output, but doesn't detail aspects like performance limits, error handling, or output format specifics beyond what the schema covers. For a tool with no annotations, this leaves significant behavioral gaps.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise and front-loaded, consisting of a single, clear phrase: 'Side-by-side text and code diff comparison.' Every word earns its place by conveying the core functionality without unnecessary elaboration, making it efficient for quick understanding.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (5 parameters, no output schema, no annotations), the description is minimally adequate. It states the purpose but lacks details on output behavior, error cases, or integration context. Without annotations or an output schema, the description doesn't fully compensate for these gaps, making it just sufficient for basic use.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The schema description coverage is 100%, with all parameters well-documented in the schema itself. The description adds no additional meaning beyond the schema, such as explaining the significance of 'left' vs 'right' or use cases for different formats. Since the schema does the heavy lifting, the baseline score of 3 is appropriate.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Side-by-side text and code diff comparison.' It specifies the verb ('diff comparison') and resource ('text and code'), making it understandable. However, it doesn't explicitly differentiate from sibling tools like 'similarity-checker' or 'code-formatter', which might have overlapping functions, preventing a perfect score.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It doesn't mention any specific contexts, prerequisites, or exclusions, such as comparing file types or handling large inputs. With many sibling tools present, including 'similarity-checker', the lack of differentiation leaves the agent without clear usage cues.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
ec-platform-comparisonプラットフォーム手数料比較BInspect
同じ商品を各ECで売った場合の利益を一覧比較 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full disclosure burden. It adds value by noting this is a 'Browser-based tool', explaining the empty parameter schema (interactive UI expected). However, it omits mutation safety, authentication requirements, or data persistence details that annotations would typically cover.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single efficient sentence front-loaded with the core value proposition (profit comparison). The parenthetical '(Browser-based tool)' is slightly abrupt but provides essential context for the empty parameter schema without excessive verbosity.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given zero parameters and no output schema, the description adequately covers the tool's purpose but leaves interaction ambiguity (how users input product details if not via parameters). For a tool with numerous sibling calculators, it should explicitly clarify why this comparison tool versus specific calculators.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema contains zero parameters. Per evaluation rules, 0 parameters establishes a baseline score of 4. The description does not add input semantics, but none are required given the schema state.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool compares profits when selling the same product across various EC platforms (各ECで売った場合の利益を一覧比較), using specific verbs (比較) and scope. It implicitly distinguishes from single-platform calculators like amazon-fba-calculator via 'various EC platforms', though it does not explicitly contrast with sibling alternatives.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no explicit guidance on when to use this comparative tool versus specific platform fee calculators (e.g., mercari-calculator, rakuten-fee-calculator) or what prerequisites exist. The user must infer usage solely from the purpose statement.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
ec-template-generatorEC定型文ジェネレーターBInspect
購入お礼/発送通知/返品対応の定型文をプラットフォーム別に生成 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full disclosure burden. The parenthetical '(Browser-based tool)' hints at execution context but does not explain output format, side effects, whether it creates persistent resources, or required authentication. For a mutation/content-generation tool with zero annotation coverage, this is insufficient behavioral disclosure.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is appropriately compact as a single sentence with the core value proposition front-loaded. The parenthetical '(Browser-based tool)' at the end is slightly cryptic (unclear if it refers to execution environment or target platform type) but does not significantly detract from readability.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given zero parameters and no output schema, the description explains what content is generated (platform-specific templates for three scenarios) but leaves critical invocation context ambiguous. Without parameters to guide the AI on which specific template to generate (purchase vs shipping vs returns) or platform selection, the description is incomplete regarding how the tool determines what to output.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema contains zero parameters. Per the evaluation rules, zero parameters warrants a baseline score of 4. There are no parameters requiring semantic elaboration beyond what the empty schema conveys.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly specifies the tool generates fixed phrases (定型文) for specific EC scenarios: purchase thank-you notes, shipping notifications, and return handling. It also mentions platform-specific generation (プラットフォーム別), distinguishing it from siblings like review-reply-template. However, it could clarify what 'platform' means (e.g., Amazon, Shopify) given the presence of specific platform calculators in siblings.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No explicit guidelines are provided for when to use this tool versus alternatives like review-reply-template or product-description-generator. While the three use cases (purchase/shipping/returns) imply the scope, there is no guidance on selection criteria, prerequisites, or when NOT to use it.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
estimate-to-expense見積書→経費明細変換BInspect
見積書PDFから経費明細表を自動生成 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It mentions 'Browser-based tool', which hints at execution environment, but lacks critical behavioral details: how PDFs are processed (e.g., upload method, supported formats), whether data is stored or transient, error handling, or output format. For a tool with zero annotation coverage, this leaves significant gaps in understanding its operation.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise: one sentence in Japanese followed by a brief English note. It's front-loaded with the main function and wastes no words. Every part ('見積書PDFから経費明細表を自動生成' and 'Browser-based tool') adds value without fluff.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (PDF processing with no parameters) and lack of annotations/output schema, the description is minimally adequate. It states what the tool does and the execution context, but lacks details on input handling (e.g., PDF requirements), output specifics, or error conditions. Without structured fields to rely on, it leaves the agent with incomplete operational context.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The tool has 0 parameters with 100% schema description coverage (empty schema). The description doesn't need to explain parameters, as there are none. It appropriately focuses on the tool's core function without redundant parameter details, meeting the baseline for zero-parameter tools.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: '見積書PDFから経費明細表を自動生成' (automatically generate expense details from estimate PDF). It specifies the verb (generate), resource (estimate PDF), and output (expense details table). However, it doesn't distinguish from siblings like 'expense-report' or 'receipt-generator', which might handle similar financial documents.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides minimal usage guidance. It mentions 'Browser-based tool', implying it runs in a browser context, but offers no explicit when-to-use rules, prerequisites, or alternatives. There's no comparison to sibling tools like 'expense-report' or 'receipt-generator' that might handle expense-related tasks differently.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
exif-removerEXIF RemoverBInspect
Strip GPS location, camera info, and metadata from photos. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It states the tool strips metadata, implying a mutation operation, but fails to describe critical traits such as whether the process is reversible, what permissions or inputs are required, how it handles errors, or the output format. The 'browser-based' hint adds minimal context, but overall, it's insufficient for a tool that modifies data.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise and front-loaded, using only one sentence that efficiently conveys the core action and context. Every word earns its place, with no wasted text, making it easy for an agent to parse quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (a mutation operation with no annotations and no output schema), the description is incomplete. It lacks details on behavioral aspects like safety, output format, or error handling. While concise, it doesn't provide enough context for an agent to reliably invoke the tool, especially compared to more descriptive siblings in the list.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description appropriately avoids discussing parameters, focusing instead on the tool's function. This meets the baseline of 4 for tools with no parameters, as it doesn't add unnecessary details beyond what the schema provides.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with specific verbs ('Strip') and resources ('GPS location, camera info, and metadata from photos'), making it easy to understand what it does. It distinguishes itself from siblings like 'pdf-metadata-remover' by specifying it's for photos and browser-based, though it doesn't explicitly contrast with all siblings. It avoids tautology by not just restating the name/title.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It mentions it's 'browser-based', which gives some context, but lacks explicit instructions on prerequisites, when-not-to-use scenarios, or comparisons to similar tools like 'pdf-metadata-remover' or 'image-crop'. This leaves the agent without clear usage direction.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
expense-reportExpense ReportBInspect
Category-based expense reports with PDF output. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It mentions 'PDF output' and 'browser-based,' but lacks critical behavioral details: whether it's read-only or mutative, if it requires authentication, any rate limits, or what happens during execution (e.g., data processing steps). For a tool with zero annotation coverage, this is insufficient.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is highly concise and front-loaded: the first part states the core functionality, and the second adds a useful constraint ('browser-based'). Both sentences earn their place by providing essential information without waste, making it easy to scan and understand.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (likely moderate, involving report generation), no annotations, no output schema, and 0 parameters, the description is minimally adequate. It covers the purpose and output format but misses behavioral transparency and usage guidelines. For a tool with no structured data to rely on, it should do more to explain how it works and when to use it.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description adds value by implying the tool might use categories for expense reports, though it doesn't specify how these are provided (e.g., via UI or defaults). Since there are no parameters, a baseline of 4 is appropriate, as the description provides some context without redundancy.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: generating 'category-based expense reports with PDF output.' It specifies the resource (expense reports) and output format (PDF), and distinguishes it from siblings like 'invoice-generator' or 'receipt-generator' by focusing on expense categorization. However, it doesn't explicitly mention the verb (e.g., 'generate' or 'create'), which slightly reduces specificity.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides minimal guidance: it notes the tool is 'browser-based,' implying it might require a browser environment, but offers no explicit when-to-use rules, alternatives, or exclusions. For example, it doesn't clarify if this is for personal vs. business expenses or how it differs from similar tools like 'profit-loss' or 'cash-flow-statement.'
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
eyecatch-makerアイキャッチメーカーAInspect
テンプレート選択→テキスト入力→ブログ用画像を即生成 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Discloses critical behavioral trait '(Browser-based tool)' indicating interactive UI flow rather than API return, and reveals 3-step workflow. However, missing output handling details (download vs URL vs inline) and persistence since no annotations cover safety/destructive hints.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely efficient: workflow arrows convey process sequence, parenthetical clarifies runtime environment. Every element earns its place with zero redundancy across 15 words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Appropriate for a zero-parameter interactive tool. Covers initiation behavior and domain, though brief mention of output format (PNG/JPG) or return behavior would strengthen completeness given lack of output schema.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Zero parameters present, meeting baseline score of 4. Description implies configuration happens via browser UI (template selection, text input), compensating for empty schema by explaining how the tool accepts input despite formal parameter count of zero.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Specific verb '生成' (generate), clear resource 'ブログ用画像' (blog eye-catch images), and distinct workflow 'テンプレート選択→テキスト入力→即生成' distinguishes it from generic image tools like image-resizer or collage-maker in the sibling list.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Implies usage scope (blog images) and workflow, but lacks explicit when-to-use guidance versus siblings like stamp-maker or image-collage, and no mention of prerequisites or alternatives.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
format-converterFormat ConverterCInspect
Convert between 11 image formats including HEIC, WebP, PNG, JPG, AVIF.
| Name | Required | Description | Default |
|---|---|---|---|
| image | Yes | Base64-encoded image to convert | |
| quality | No | Output quality (1-100, for lossy formats) | |
| outputFormat | No | Target output format | png |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions the conversion action but doesn't describe side effects (e.g., whether the original image is preserved), performance traits (e.g., speed or size limits), or error conditions (e.g., unsupported formats). For a tool with no annotations, this leaves significant gaps in understanding its behavior.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence with zero waste. It front-loads the core purpose ('Convert between 11 image formats') and provides relevant examples without unnecessary elaboration. Every word earns its place, making it highly concise and well-structured.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (format conversion with quality settings) and lack of annotations or output schema, the description is insufficient. It doesn't explain what the tool returns (e.g., base64 output, error messages), how to handle unsupported formats, or the impact of quality settings. For a tool with no structured behavioral data, more descriptive context is needed.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema fully documents the three parameters (image, quality, outputFormat). The description adds no parameter-specific information beyond implying format conversion, which is already covered by the schema's enum for outputFormat. This meets the baseline for high schema coverage, but doesn't enhance understanding of parameters.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: converting between image formats, listing specific examples (HEIC, WebP, PNG, JPG, AVIF). It uses a specific verb ('Convert') and identifies the resource ('image formats'). However, it doesn't explicitly differentiate from sibling tools like 'image-resizer' or 'image-to-pdf', which have related but distinct purposes.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It doesn't mention prerequisites (e.g., needing an image input), exclusions (e.g., not for non-image files), or comparisons to siblings like 'image-to-pdf' or 'base64-converter'. Usage is implied by the tool's name and description but not explicitly stated.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
freelance-tax-calcフリーランス手取りシミュレーターAInspect
所得税・住民税・事業税・国保・年金・消費税を全て計算 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided. Description adds 'Browser-based tool' indicating interactive web interface rather than API calculation, but omits safety profile (read-only), whether it stores data, or whatoutput format users receive.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Single efficient sentence front-loaded with tax scope followed by implementation hint. No redundant text.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Lists six complex tax types but lacks output schema description. Given complexity of Japanese freelance tax stack, could clarify that 'browser-based' implies user interaction required to view results, not automatic return.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Zero parameters per input schema. Baseline 4 applies as no parameter documentation burden exists.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Lists specific tax categories (所得税・住民税・事業税・国保・年金・消費税) distinguishing freelance-specific obligations (business tax, consumption tax) from employee-focused siblings like take-home-pay-calc. Title clarifies 'freelance' scope.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance on when to select this versus siblings like tax-return-calc (filing), salary-vs-freelance (comparison), or nhi-calc/social-insurance-calc (single-tax calculators). No prerequisites or constraints mentioned.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
furigana-adderFurigana AdderBInspect
Auto-add reading aids to kanji by grade level. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It mentions 'Auto-add' and 'Browser-based tool', hinting at automation and platform, but fails to disclose critical behaviors like input/output format, permissions, rate limits, or error handling. This leaves significant gaps for a tool with potential processing implications.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise with two brief phrases: 'Auto-add reading aids to kanji by grade level. (Browser-based tool).' Every word contributes meaning, and it's front-loaded with the core purpose. No wasted sentences.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given 0 parameters and no output schema, the description is minimally adequate but incomplete. It lacks details on how the tool operates (e.g., input method, output format), behavioral traits, and usage context, which are crucial for a tool that modifies text. The absence of annotations exacerbates this gap.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description adds context about 'grade level' and 'Browser-based tool', which provides useful semantic information beyond the empty schema, though it's not strictly about parameters.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Auto-add reading aids to kanji by grade level.' It specifies the action (add), resource (reading aids/kanji), and scope (by grade level). However, it doesn't explicitly differentiate from siblings, which are unrelated tools (e.g., calculators, generators, converters) with no direct alternatives for furigana addition.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides minimal guidance with '(Browser-based tool)', implying it's for web use, but lacks explicit when-to-use instructions, prerequisites, or alternatives. No context on when to choose this over manual methods or other tools is given.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
furusato-tax-calcふるさと納税上限額シミュレーターBInspect
自己負担2,000円で済む控除上限額を瞬時に逆算 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It successfully discloses the browser-based nature and the specific calculation methodology (2000 yen self-burden). However, it lacks critical behavioral details: it does not specify the output format (HTML, JSON, text), whether the tool is idempotent, if it stores user data, or what interactive elements the browser interface contains.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise—one sentence with a parenthetical qualifier. The core functionality (reverse calculation of deduction limit) is front-loaded, followed by the specific constraint (2000 yen burden) and the implementation hint. There is no redundancy or extraneous text.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool has zero parameters, no annotations, and no output schema, the description is minimally adequate but incomplete. It identifies what is calculated but not what the user receives (a URL, HTML content, or raw numbers). For a 'calculator' with no input parameters, it should explicitly state that it launches an interactive web interface or returns a form, rather than assuming the 'Browser-based' parenthetical suffices.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema contains zero parameters, establishing a baseline score of 4. The description implicitly explains why parameters might not be needed (it describes a calculation target rather than inputs), referencing the 'browser-based' nature which suggests an interactive UI handles inputs. No additional parameter documentation is required or provided.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly identifies the specific action (逆算/reverse calculation) and resource (控除上限額/deduction upper limit) for the Furusato Nozei (hometown tax) system. It specifies the unique constraint of '自己負担2,000円' (2,000 yen self-burden), which defines the calculation scope. However, it does not explicitly distinguish this from sibling tax calculators like 'tax-return-calc' or 'resident-tax-calc', though the domain specificity is implied.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides minimal usage guidance. While the 'Browser-based tool' parenthetical hints at the interaction model, there is no explicit guidance on when to select this tool versus the numerous sibling tax calculators (e.g., 'freelance-tax-calc', 'withholding-tax-calc'). It does not state prerequisites (e.g., required income data) or when not to use it.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
grade-calculatorGrade CalculatorCInspect
Convert 3-aspect evaluations to ABC and 5-level grades.
| Name | Required | Description | Default |
|---|---|---|---|
| aThreshold | No | Threshold for A grade (%) | |
| bThreshold | No | Threshold for B grade (%) | |
| attitudeMax | No | Attitude max score | |
| thinkingMax | No | Thinking & Expression max score | |
| knowledgeMax | No | Knowledge & Skills max score | |
| attitudeScore | Yes | Attitude score (主体的態度の得点) | |
| thinkingScore | Yes | Thinking & Expression score (思考・判断・表現の得点) | |
| knowledgeScore | Yes | Knowledge & Skills score (知識・技能の得点) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It states the conversion function but doesn't explain how the conversion works (e.g., formula, thresholds), what the output looks like, or any error handling. For a tool with 8 parameters and no annotations, this is insufficient.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence with zero waste. It's front-loaded and directly states the tool's function without unnecessary details, making it easy to parse quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity (8 parameters, no annotations, no output schema), the description is incomplete. It doesn't explain the conversion logic, output format, or behavioral traits like error handling. For a calculation tool with multiple inputs and no structured output, more context is needed for effective use.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema fully documents all parameters with descriptions and defaults. The description adds no additional parameter semantics beyond implying three input scores (knowledge, thinking, attitude) are used, which is already covered in the schema. Baseline 3 is appropriate when schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: converting three evaluation aspects into two grade systems (ABC and 5-level). It specifies the input (3-aspect evaluations) and output (ABC and 5-level grades), making the verb+resource explicit. However, it doesn't distinguish from siblings, as no similar grading tools are listed among siblings.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives. The description lacks context about prerequisites, typical scenarios, or exclusions. It doesn't mention any sibling tools for comparison, leaving usage entirely implicit.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
hash-generatorHash GeneratorBInspect
Compute MD5, SHA-256, and other hash values.
| Name | Required | Description | Default |
|---|---|---|---|
| text | Yes | Text to hash | |
| algorithm | No | Hash algorithm | SHA-256 |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It states the tool computes hash values but doesn't mention any behavioral traits such as performance characteristics, error handling, or output format (e.g., hex string, base64). For a tool with no annotations, this leaves significant gaps in understanding how it behaves.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise—a single sentence—with no wasted words. It front-loads the core purpose ('Compute MD5, SHA-256, and other hash values'), making it easy to scan and understand quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's low complexity (two parameters, no nested objects) and high schema coverage, the description is minimally adequate. However, with no output schema and no annotations, it lacks details on return values (e.g., hash format) and behavioral context, which could be important for an agent invoking this tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, with clear descriptions for both parameters ('text' and 'algorithm'). The description adds minimal value beyond the schema, only implying that multiple algorithms are supported. Since the schema already documents parameters thoroughly, the baseline score of 3 is appropriate.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Compute MD5, SHA-256, and other hash values.' It specifies the verb ('Compute') and resource ('hash values'), and lists example algorithms. However, it doesn't explicitly differentiate from sibling tools like 'uuid-generator' or 'base64-converter', which are also data transformation utilities but for different purposes.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It doesn't mention any prerequisites, context for hashing (e.g., security, data integrity), or comparisons with sibling tools like 'base64-converter' (encoding) or 'uuid-generator' (unique identifiers), leaving the agent to infer usage from the name alone.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
hashtag-generatorハッシュタグジェネレーターAInspect
キーワードからSNS用ハッシュタグを自動生成。TikTok/Instagram/X対応 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description carries full burden. Discloses critical 'Browser-based tool' trait indicating interactive/URL-opening execution, but omits what the tool returns (list of hashtags? opens browser window?) and any side effects.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two highly efficient sentences with action front-loaded. 'Browser-based tool' parenthetical earns its place by disclosing execution environment. Zero waste.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Adequate for a simple 0-param tool, but lacks description of return values or output behavior (critical given no output schema exists). 'Browser-based' hints at interaction model but could clarify what the agent receives back.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema has 0 parameters, setting baseline at 4. Description mentions 'キーワードから' (from keywords), adding semantic context that the browser-based tool expects keyword input from the user, even though the agent passes no parameters.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Specific verb (自動生成/auto-generate), clear resource (SNS用ハッシュタグ/SNS hashtags), and precise scope (TikTok/Instagram/X). Effectively distinguishes from sibling youtube-tag-generator (platform-specific) and sns-post-template (post templates vs hashtags).
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Lists supported platforms (TikTok/Instagram/X) providing implicit context for when to use (SNS marketing), but lacks explicit guidance on when to choose this over siblings like youtube-tag-generator or meta-description-generator.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
html-sanitizerHTML SanitizerBInspect
Sanitize HTML removing dangerous tags and attributes.
| Name | Required | Description | Default |
|---|---|---|---|
| html | Yes | HTML string to sanitize | |
| level | No | Sanitization level | moderate |
| removeEmpty | No | Remove empty elements |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions 'removing dangerous tags and attributes,' which implies a read-only or transformation operation, but fails to detail critical aspects like error handling, performance implications, security guarantees, or what constitutes 'dangerous.' For a security-related tool with zero annotation coverage, this is a significant gap in transparency.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence with zero waste—it directly states the tool's purpose without unnecessary words. It is appropriately sized and front-loaded, making it easy for an agent to parse quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (HTML sanitization with three parameters), no annotations, and no output schema, the description is minimally adequate. It covers the basic purpose but lacks details on behavior, output format, or error handling. It meets the minimum viable threshold but has clear gaps in providing a complete context for safe and effective use.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all three parameters (html, level, removeEmpty) with descriptions. The description adds no additional meaning beyond what the schema provides, such as examples of sanitization levels or the impact of 'removeEmpty.' Baseline score of 3 is appropriate when the schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the verb ('sanitize') and resource ('HTML'), specifying the action of removing dangerous tags and attributes. It distinguishes itself from other tools in the sibling list, which are unrelated to HTML processing. However, it doesn't explicitly differentiate from potential similar tools like 'html-formatter' or 'html-validator' that might exist elsewhere, keeping it from a perfect score.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives, prerequisites, or exclusions. It lacks context such as typical use cases (e.g., user input sanitization, content rendering) or comparisons with other tools in the sibling list, which are unrelated. This leaves the agent without explicit usage direction.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
http-probeHTTP ProbeAInspect
Probe a URL and return status code, headers, response time, and redirect info.
| Name | Required | Description | Default |
|---|---|---|---|
| url | Yes | URL to probe | |
| method | No | HTTP method (HEAD is faster, GET gets body info) | HEAD |
| timeout | No | Timeout in milliseconds | |
| followRedirects | No | Follow redirects (false=show redirect target) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description carries full burden. It effectively discloses output composition (four specific data points returned) but omits safety profile, error handling (timeouts, DNS failures), and side-effect risks.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Single 12-word sentence with zero waste. Front-loaded with action ('Probe') and immediately followed by return value specification. Every word earns its place.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
No output schema exists, but description compensates by enumerating the four specific return components. Adequate for a simple 4-parameter probe tool, though error case documentation would improve it.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema has 100% description coverage (url, method, timeout, followRedirects all documented), establishing baseline 3. Description mentions 'redirect info' which semantically reinforces the followRedirects parameter's purpose, but adds no explicit parameter guidance.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Specific verb 'Probe' + resource 'URL' + detailed outputs (status code, headers, response time, redirect info) clearly distinguishes this diagnostic tool from siblings (calculators, generators, converters).
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No explicit when-to-use guidance or comparison to alternatives (e.g., og-preview which also likely fetches URLs). Description only states what it does, not when to choose it.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
id-photo-makerID Photo MakerAInspect
Create passport/ID photos without uploading to any server. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It discloses key behavioral traits: the tool creates photos (implying a write operation) and operates locally without server uploads (privacy/security context). However, it lacks details on output format, quality, supported image types, or any limitations (e.g., browser compatibility, processing time).
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise (two short phrases) and front-loaded with the core purpose. Every word earns its place: 'Create passport/ID photos' states the action, and 'without uploading to any server. (Browser-based tool)' adds critical constraints efficiently. There is no wasted text.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (a photo creation tool with no parameters), the description is adequate but has gaps. It lacks an output schema, and without annotations, it doesn't fully cover behavioral aspects like output format or error handling. However, for a zero-parameter tool, it provides enough context to understand the basic purpose and privacy feature.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description adds value by clarifying the tool's scope (passport/ID photos) and operational context (browser-based, no server uploads), which goes beyond the empty schema. A baseline of 4 is appropriate as it compensates for the lack of parameters with meaningful context.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Create passport/ID photos' specifies the verb (create) and resource (photos), and 'without uploading to any server' adds a key constraint. However, it doesn't explicitly differentiate from sibling tools like 'product-photo-studio' or 'thumbnail-ctr-predictor', which might also involve image creation or manipulation.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage context through 'without uploading to any server' and 'Browser-based tool', suggesting privacy and local processing benefits. However, it doesn't provide explicit guidance on when to use this tool versus alternatives (e.g., 'image-crop' or 'bg-remover' for other photo edits), nor does it mention any prerequisites or exclusions.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
image-brightnessBrightness AdjustBInspect
AI auto-correction for dark or underexposed photos. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden for behavioral disclosure. It mentions 'AI auto-correction' and 'Browser-based tool,' hinting at automated processing and execution environment, but lacks critical details: whether the tool modifies the original image, what output format to expect, any rate limits, or authentication needs. For a tool that likely alters images, this is insufficient.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise: two short phrases that directly state the tool's function and technical context. Every word earns its place, with no redundancy or fluff. It's front-loaded with the core purpose, making it easy to scan.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (likely image processing with AI), lack of annotations, and no output schema, the description is incomplete. It misses key details: what the tool returns (e.g., a modified image file, a brightness score), how it handles errors, or any side effects. The agent lacks sufficient context to use this tool effectively beyond a vague understanding.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description doesn't add parameter details, which is appropriate. A baseline of 4 is applied since the schema fully covers the zero parameters, and the description doesn't introduce confusion.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'AI auto-correction for dark or underexposed photos.' It specifies the action (auto-correction) and target resource (photos with brightness issues). However, it doesn't explicitly differentiate from sibling tools like 'image-color-picker' or 'image-crop' beyond the brightness focus, preventing a perfect score.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides minimal usage guidance: it implies use for 'dark or underexposed photos' but offers no explicit when-to-use rules, prerequisites, or alternatives. The parenthetical '(Browser-based tool)' adds a technical constraint but doesn't help the agent choose between this and other image tools like 'bg-remover' or 'image-resizer'.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
image-collageImage CollageBInspect
Combine multiple photos into a single image with layout templates. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It lacks critical behavioral details: it doesn't specify input methods (e.g., file uploads, URLs), supported image formats, output format (e.g., PNG, JPEG), size limits, or whether it's a read-only or mutating operation. The 'Browser-based' note is helpful but insufficient for a tool with potential complexity.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is concise and front-loaded: the first sentence clearly states the core functionality, and the second adds useful context ('Browser-based'). There's no wasted text, though it could be slightly more informative without losing efficiency.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given 0 parameters and no output schema, the description is minimally complete. It explains what the tool does but lacks details on behavior, inputs, and outputs. For a tool that likely involves image processing and user interaction, more context on limitations or usage would improve completeness.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description doesn't add param info, which is appropriate. Baseline is 4 for 0 params, as it avoids redundancy and focuses on the tool's purpose.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Combine multiple photos into a single image with layout templates.' It specifies the verb ('Combine'), resource ('multiple photos'), and output ('single image'), and distinguishes it from siblings like 'image-crop' or 'image-resizer'. However, it doesn't explicitly differentiate from all image-related tools (e.g., 'image-to-pdf'), so it's not a perfect 5.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides minimal usage guidance. It mentions the tool is 'Browser-based', which hints at execution context, but offers no explicit when-to-use rules, prerequisites, or alternatives. For instance, it doesn't clarify if it's for creating collages from uploaded files versus URLs, or when to choose this over other image manipulation tools.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
image-color-picker画像カラーピッカーAInspect
画像から色を瞬時に特定。スポイト+自動パレット抽出 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations, the description carries full burden. It successfully discloses the browser-based nature and interactive eyedropper mechanism, but omits critical behavioral details like output format (HEX/RGB), persistence of extracted colors, or file size limitations.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely efficient two-part structure: first sentence states core function, second specifies mechanism and runtime environment. No redundant padding despite bilingual presentation.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a zero-parameter tool with no output schema, the description covers the essential interaction model. However, given the server's extensive image-tool ecosystem (bg-remover, image-crop, etc.), it should clarify how images are fed into this browser tool (upload vs URL vs current view).
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema has zero parameters, triggering the baseline score of 4. The description appropriately doesn't invent parameter documentation, though it implies image input is handled through the browser interface rather than schema parameters.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description clearly states the tool identifies/extracts colors from images using an eyedropper and automatic palette extraction. The phrase '画像から' (from image) and 'スポイト' (eyedropper) specifically distinguishes it from siblings like 'color-palette' (theoretical generation) and 'color-converter' (format conversion).
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
While '(Browser-based tool)' hints at the interaction model, the description lacks explicit guidance on when to choose this over 'color-palette' or 'image-resizer', and doesn't explain prerequisites like image upload mechanisms despite having zero input parameters.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
image-cropImage CropBInspect
Drag-to-select cropping with preset aspect ratios. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions 'Browser-based tool', indicating it runs in a browser, but doesn't describe what the tool actually does behaviorally—such as whether it modifies images in-place, requires user interaction, or has any side effects. For a tool with zero annotation coverage, this is a significant gap in transparency.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise—two short phrases separated by a period. Every word earns its place: 'Drag-to-select cropping' defines the core action, 'with preset aspect ratios' adds a key feature, and '(Browser-based tool)' provides essential context. It's front-loaded with the main purpose and wastes no space.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool has 0 parameters, no annotations, and no output schema, the description is minimally adequate. It states what the tool does and its environment, but lacks details on behavior, output, or integration with siblings. For a tool with such low complexity, it's complete enough to be functional but leaves gaps in understanding how it operates or what it returns.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, meaning no parameters are documented in the schema. The description doesn't add parameter details, but since there are no parameters, the baseline is 4. It implicitly suggests parameters might be handled via UI (drag-to-select), but doesn't explain semantics beyond that.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's function: 'Drag-to-select cropping with preset aspect ratios.' It specifies the action (drag-to-select cropping) and key feature (preset aspect ratios), which distinguishes it from generic image tools. However, it doesn't explicitly differentiate from sibling tools like 'image-resizer' or 'smart-resize', which is why it's not a 5.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides minimal usage guidance. It mentions 'Browser-based tool', which implies it requires a browser environment, but doesn't specify when to use this tool versus alternatives (e.g., 'image-resizer' for resizing without cropping, 'smart-resize' for automated adjustments). No explicit when/when-not or alternative recommendations are included.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
image-resizerImage ResizerCInspect
Resize images to exact pixel dimensions with aspect ratio lock.
| Name | Required | Description | Default |
|---|---|---|---|
| fit | No | Resize fit mode (contain=fit within, cover=fill & crop, fill=stretch) | contain |
| image | Yes | Base64-encoded image (PNG, JPEG, WebP, AVIF, TIFF, GIF supported) | |
| width | No | Target width in pixels | |
| format | No | Output image format | png |
| height | No | Target height in pixels | |
| quality | No | Output quality (1-100, for lossy formats) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions 'aspect ratio lock' but fails to describe critical behaviors like whether the operation is destructive (e.g., overwrites original), performance characteristics (e.g., processing time), error handling, or output specifics. This leaves significant gaps for a mutation tool.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence with zero waste. It is front-loaded with the core purpose and avoids unnecessary elaboration, making it highly concise and well-structured.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (6 parameters, mutation operation) and lack of annotations and output schema, the description is insufficient. It omits behavioral details (e.g., side effects, error cases) and output information, failing to provide a complete context for safe and effective use.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, providing detailed parameter documentation. The description adds no additional parameter semantics beyond what's in the schema, such as explaining interactions between 'fit' and 'aspect ratio lock'. Baseline 3 is appropriate since the schema handles the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with a specific verb ('Resize') and resource ('images'), and specifies 'exact pixel dimensions with aspect ratio lock' to clarify the operation. However, it does not explicitly differentiate from sibling tools like 'image-crop', 'smart-resize', or 'upscaler', which might have overlapping functionality, so it doesn't fully achieve sibling differentiation.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It lacks any mention of prerequisites, exclusions, or comparisons with sibling tools such as 'image-crop' or 'smart-resize', leaving the agent without context for selection.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
image-to-pdfImage to PDFBInspect
Combine multiple images into a single PDF document. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It states the tool combines images into a PDF and is browser-based, but lacks details on permissions, rate limits, file size constraints, supported image formats, or output behavior. For a tool with zero annotation coverage, this is insufficient for safe and effective use.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise and front-loaded: it states the core function in the first clause and adds a contextual note in parentheses. Every word earns its place, with no redundancy or unnecessary elaboration.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (simple conversion with no parameters) and lack of annotations or output schema, the description is minimally adequate. It explains what the tool does but lacks behavioral details (e.g., supported formats, limitations) and output information. It meets the bare minimum for a zero-parameter tool but leaves gaps in operational context.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description does not add parameter details, which is appropriate. A baseline score of 4 is given because the schema fully covers the parameters (none), and the description does not need to compensate.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Combine multiple images into a single PDF document.' It specifies the verb ('Combine'), resource ('images'), and outcome ('single PDF document'), making the function unambiguous. However, it does not explicitly differentiate from sibling tools like 'pdf-join' or 'pdf-to-image', which prevents a perfect score.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides minimal guidance: it mentions the tool is 'Browser-based,' which hints at usage context but does not specify when to use this tool versus alternatives (e.g., 'pdf-join' for merging PDFs or 'image-collage' for combining images without PDF conversion). No explicit when/when-not or alternative recommendations are included.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
internal-link-map内部リンクマップBInspect
記事一覧から内部リンク構造を可視化。収益記事への導線を最適化 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full disclosure burden. It adds 'browser-based tool' indicating execution context, but fails to clarify if this modifies site data (destructive vs read-only), required permissions, output format, or side effects like opening browser windows.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely concise two-sentence structure with zero waste: function sentence, benefit sentence, and technical parenthetical. Information is front-loaded with the core action (visualize) stated immediately.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given zero input parameters and no output schema/annotations, the description adequately explains the tool's purpose but lacks crucial execution context. It should specify whether it returns visualization data, opens a browser UI, or modifies the site's link structure.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema has zero parameters, establishing baseline score of 4. The description correctly implies no user-supplied parameters are needed (works from internal article list), requiring no additional parameter documentation.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool visualizes internal link structures from article lists and optimizes flow to revenue articles (specific verb+resource). However, the meaning of 'browser-based tool' is slightly ambiguous (opens browser vs returns HTML), and it doesn't explicitly distinguish from potential SEO/analysis siblings.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No explicit guidance on when to use versus alternatives (e.g., manual sitemap analysis), prerequisites (must have articles indexed?), or when-not-to-use. The implied usage relies solely on understanding the value proposition.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
investment-simulatorInvestment SimulatorBInspect
NISA/iDeCo compound growth simulation.
| Name | Required | Description | Default |
|---|---|---|---|
| mode | No | Simulation mode | tsumitate-nisa |
| inflationRate | No | Inflation rate (%) | |
| monthlyAmount | Yes | Monthly investment amount in JPY | |
| initialLumpSum | No | Initial lump sum investment in JPY | |
| investmentYears | No | Investment period in years | |
| annualReturnRate | No | Annual return rate (%) | |
| enableInflationAdjustment | No | Enable inflation adjustment |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It mentions 'simulation' which implies read-only/non-destructive behavior, but doesn't clarify if it's a calculation, projection, or what outputs to expect. It lacks details on rate limits, error conditions, or whether it requires authentication. The description is too brief to provide meaningful behavioral context beyond the basic purpose.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise—a single phrase with no wasted words. It's front-loaded with the core purpose. For a tool with comprehensive schema documentation, this brevity is appropriate and efficient.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (7 parameters, no output schema, no annotations), the description is minimal but adequate as a starting point. The schema provides full parameter documentation, but the description doesn't address output format, error handling, or usage context. It meets minimum viability but leaves gaps in behavioral understanding.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so parameters are well-documented in the schema itself. The description adds no additional parameter semantics beyond implying the tool handles NISA/iDeCo modes and compound growth. It doesn't explain parameter interactions, default behaviors, or provide examples. Baseline score of 3 is appropriate since the schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description 'NISA/iDeCo compound growth simulation' clearly states the tool's purpose as simulating compound growth for specific Japanese investment accounts (NISA/iDeCo). It uses specific terms ('compound growth simulation') and identifies the target resources, but doesn't explicitly differentiate from sibling tools, which appear to be various calculators and generators without direct overlap in investment simulation.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It doesn't mention prerequisites, typical use cases, or compare it to other investment calculators in the sibling list (e.g., 'npv-irr-calc', 'bond-yield-calc'). The user must infer usage from the tool name and parameters alone.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
invoice-generatorInvoice GeneratorBInspect
Japan qualified invoice compliant. PDF output in 3 seconds. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden. It discloses that the tool is 'Browser-based' and outputs PDFs in '3 seconds', adding useful behavioral context about performance and execution. However, it doesn't cover aspects like error handling, authentication needs, or rate limits, leaving gaps for a mutation tool.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise with three short phrases, front-loading key information (compliance, output format, speed, and environment). Every sentence earns its place without redundancy.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no annotations, no output schema, and 0 parameters, the description provides basic purpose and behavior but lacks details on usage context, error cases, or output specifics. It's minimally adequate for a simple tool but has clear gaps in guiding effective use.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description doesn't add param details, which is appropriate, but it implies the tool generates invoices without specifying input requirements, slightly reducing clarity. Baseline is 4 for zero parameters.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description states the tool generates invoices that are 'Japan qualified invoice compliant' with 'PDF output', which clarifies its purpose beyond the name/title. However, it doesn't specify what inputs are required or how it differs from sibling tools like 'receipt-generator' or 'purchase-order', making it somewhat vague.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives. The description mentions 'Browser-based tool', which hints at the execution environment but doesn't specify use cases, prerequisites, or exclusions compared to other tools on the server.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
journal-entryJournal EntryBInspect
Create debit/credit journal entries with PDF export. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It mentions 'Create' (implying a write operation) and 'PDF export', but doesn't disclose behavioral traits like permissions needed, whether entries are saved or temporary, error handling, or rate limits. The 'Browser-based tool' note adds some context but is vague. For a mutation tool with zero annotation coverage, this is insufficient.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise: one sentence plus a parenthetical note. Every part earns its place by stating the core functionality and an important implementation detail (browser-based). It's front-loaded with the main purpose and wastes no words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool has 0 parameters, no annotations, and no output schema, the description is minimal but covers the basic purpose. However, for a 'create' operation, it lacks details on what happens after creation (e.g., where entries are stored, if PDF is downloadable), making it incomplete for full agent understanding. It's adequate but has clear gaps in behavioral context.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameters are documented in the schema. The description doesn't mention any parameters, which is appropriate here. Since there are no parameters to explain, it meets the baseline expectation without needing to compensate for gaps.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Create debit/credit journal entries with PDF export.' It specifies the action (create), resource (journal entries), and additional capability (PDF export). However, it doesn't explicitly differentiate from sibling tools like 'balance-sheet' or 'profit-loss', which might be related financial tools, so it misses full sibling distinction.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It mentions it's 'Browser-based tool', which might imply usage context, but doesn't specify prerequisites, scenarios, or exclusions. Without explicit when/when-not instructions, it offers minimal usage direction.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
json-expertJSON FormatterCInspect
Format, validate, minify JSON with tree view.
| Name | Required | Description | Default |
|---|---|---|---|
| json | Yes | JSON string to format, validate, or minify | |
| indent | No | Indentation spaces | |
| minify | No | Minify instead of format | |
| sortKeys | No | Sort object keys alphabetically |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions 'tree view' as a feature, which adds some context beyond basic formatting, but fails to describe critical traits like error handling (e.g., validation failures), performance implications (e.g., large JSON handling), or output format details. This is inadequate for a tool with multiple functions.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise and front-loaded, listing key functions in a single, efficient phrase. Every word earns its place, with no redundant or vague language, making it easy to scan and understand quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (multiple functions like formatting, validating, minifying) and lack of annotations and output schema, the description is insufficient. It doesn't explain what 'tree view' entails, how validation errors are reported, or the format of results, leaving significant gaps for an agent to operate effectively.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema fully documents all parameters (json, indent, minify, sortKeys). The description adds no additional meaning beyond the schema, such as examples or use cases for parameters. Baseline 3 is appropriate as the schema handles parameter semantics effectively.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with specific verbs (format, validate, minify) and resource (JSON), and mentions 'tree view' as an additional feature. It distinguishes itself from siblings like 'json-schema-validator' or 'json-to-csv' by focusing on formatting operations, though it doesn't explicitly contrast with them.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It lists functions but doesn't indicate scenarios, prerequisites, or exclusions. For example, it doesn't clarify if it's for debugging, data preparation, or other contexts, leaving the agent to infer usage.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
json-schema-validatorJSON Schema ValidatorBInspect
Validate JSON data against a JSON Schema.
| Name | Required | Description | Default |
|---|---|---|---|
| json | Yes | JSON string to validate | |
| draft | No | Schema draft version (auto-detected if omitted) | |
| schema | Yes | JSON Schema string |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It only states the basic action ('validate') without detailing error handling, performance characteristics, or output format. For a tool with no annotations, this is a significant gap in transparency.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, clear sentence with zero waste. It is front-loaded and efficiently conveys the core purpose without unnecessary elaboration, making it highly concise and well-structured.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (3 parameters, no output schema, no annotations), the description is minimally adequate. It states what the tool does but lacks details on behavior, output, or usage context, leaving gaps that could hinder effective agent operation.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all parameters thoroughly. The description does not add any semantic details beyond what the schema provides, such as examples or constraints, resulting in the baseline score for high schema coverage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with a specific verb ('validate') and resource ('JSON data against a JSON Schema'), making it immediately understandable. However, it does not differentiate this tool from potential siblings, as there are no obvious validation-related tools in the sibling list, so it lacks explicit distinction.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It does not mention any context, prerequisites, or exclusions, leaving the agent to infer usage based solely on the tool's name and purpose.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
json-to-csvJSON to CSVCInspect
Convert JSON arrays of objects to CSV format.
| Name | Required | Description | Default |
|---|---|---|---|
| json | Yes | JSON array of objects to convert to CSV | |
| delimiter | No | CSV delimiter | , |
| includeHeader | No | Include header row |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description carries full burden. States conversion direction but omits critical behavioral details: handling of nested objects/arrays, character encoding, line ending format, or error behavior for invalid JSON. Does not disclose output structure beyond 'CSV'.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Single sentence, seven words. Efficiently structured and front-loaded with no redundancy. However, excessive brevity given the lack of annotations and output schema leaves significant gaps.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Input schema is well-documented (100% coverage), but description lacks behavioral context needed for a data conversion tool (flattening logic, encoding). No output schema exists to describe return format. Adequate but minimal for the tool's complexity.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema has 100% description coverage documenting all three parameters (json, delimiter, includeHeader). Description mentions 'JSON arrays of objects' which reinforces the schema but adds no additional semantic detail about parameter usage (e.g., when to choose tab vs comma delimiters).
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Clear verb 'Convert' with specific resource 'JSON arrays of objects' and target format 'CSV'. However, does not distinguish from siblings like 'csv-excel-converter' or 'json-to-typescript'.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance on when to use this tool versus alternatives like csv-excel-converter or xml-json-converter. No mention of prerequisites (e.g., valid JSON structure) or error conditions.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
json-to-typescriptJSON to TypeScriptBInspect
Infer TypeScript interface definitions from JSON samples.
| Name | Required | Description | Default |
|---|---|---|---|
| json | Yes | JSON sample to infer TypeScript types from | |
| rootName | No | Name for the root interface | Root |
| exportTypes | No | Add export keyword to interfaces |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description carries full burden. The verb 'Infer' correctly suggests generative/deductive behavior rather than simple conversion, hinting that the tool analyzes structure to generate types. However, lacks disclosure of output format (code string vs file), error handling for invalid JSON, or whether nested objects generate additional interfaces.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Single 9-word sentence with zero redundancy. Front-loaded with active verb. Given 100% schema coverage, this level of brevity is appropriate—every word earns its place without requiring additional sentences.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Minimal viable description for a simple 3-parameter conversion tool. Adequate given flat schema structure and complete parameter documentation. Missing output specification (returns TypeScript code as string) which would be helpful given no output schema exists.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100% (all 3 parameters documented in schema), establishing baseline 3. Description adds no parameter-specific semantics beyond what schema provides (e.g., no usage examples for rootName or guidance on exportTypes defaults).
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Clear verb 'Infer' plus resource 'TypeScript interface definitions' and source 'JSON samples'. Specifies the exact output format (interfaces) distinguishing it from generic JSON converters like json-to-csv or yaml-json. However, lacks explicit differentiation from json-expert or json-schema-validator siblings.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance provided on when to use this tool versus alternatives (e.g., when to prefer over json-expert for type generation). No mention of prerequisites like valid JSON syntax or limitations on sample size/complexity.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
jwt-decoderJWT DecoderCInspect
Decode and inspect JWT token contents.
| Name | Required | Description | Default |
|---|---|---|---|
| token | Yes | JWT token to decode |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It mentions decoding and inspecting, implying a read-only operation, but fails to disclose behavioral traits like whether it validates signatures, handles errors, or returns structured output. This leaves gaps in understanding the tool's behavior.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence with zero waste. It is front-loaded and appropriately sized for a simple tool, making it easy to understand quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the lack of annotations and output schema, the description is incomplete. It does not explain what the decoded output includes (e.g., header, payload, validation status) or error handling, which is crucial for a tool with no structured output documentation.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, with the parameter 'token' clearly documented. The description adds no additional meaning beyond the schema, such as token format examples or constraints, so it meets the baseline for adequate but not enhanced parameter semantics.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the action ('decode and inspect') and resource ('JWT token contents'), making the purpose evident. However, it does not differentiate from siblings, as no similar tools (e.g., 'json-expert' or 'base64-converter') are mentioned for comparison.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives. The description lacks context about prerequisites, such as token format or validation, and does not reference sibling tools for related tasks.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
keyword-difficulty-checkerKW難易度チェッカーBInspect
キーワードの競合強度・必要文字数・上位表示の難易度を即判定 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden. It adds the 'Browser-based tool' context (indicating UI interaction) and '即判定' (instant determination) for timing, but lacks critical safety disclosures: read-only status, network requirements, rate limits, or what happens to the evaluated data.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Single sentence efficiently packs three specific evaluation criteria and the delivery mechanism. Every element earns its place; no redundancy or filler text.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a zero-parameter tool with no output schema, the description minimally covers the functional output, but leaves a significant gap regarding input mechanism (how the keyword is specified) and lacks completeness regarding behavioral traits that annotations would typically cover.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema contains zero parameters (100% coverage trivially), meeting the baseline score of 4. The description does not clarify how the keyword input is provided (likely through the browser UI given the 'browser-based' note), but with an empty schema, no additional parameter semantics are expected.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool evaluates keyword competition strength, required character count, and ranking difficulty (上位表示の難易度), specifying both the resource (keywords) and the analytical dimensions. However, it could more explicitly distinguish from siblings like 'sales-writing-analyzer' or 'meta-description-generator' by explicitly mentioning SEO context.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description lists what metrics are evaluated but provides no guidance on when to use this tool versus alternatives like 'cvr-improvement-checker' or content generation siblings. There are no explicit prerequisites, exclusions, or scenarios mentioned (e.g., 'use when planning SEO content').
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
markdown-editorMarkdownエディタBInspect
リアルタイムプレビュー付きMarkdownエディタ。GFM対応 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It mentions 'real-time preview' and 'GFM compatibility' which are useful behavioral traits, but fails to describe critical aspects like whether it saves files locally, requires authentication, has rate limits, or what the user interface looks like. For a browser-based tool with zero annotation coverage, this leaves significant gaps in understanding how it behaves.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise with just two brief phrases that each add value: 'Markdown editor with real-time preview' establishes core functionality, 'GFM compatible' adds technical specificity, and 'Browser-based tool' provides implementation context. Every word earns its place with zero wasted text.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (a browser-based editor with real-time preview), the description provides basic functionality but lacks important context. With no output schema and no annotations, it doesn't explain what the tool returns or how the editing interface works. While adequate for a simple tool, it should ideally describe the user experience more fully for an interactive editing tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The tool has 0 parameters with 100% schema description coverage, so the schema already fully documents the input requirements. The description appropriately doesn't discuss parameters since none exist. A baseline of 4 is appropriate for zero-parameter tools where the schema handles all documentation needs.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose as a 'Markdown editor with real-time preview' and specifies GFM (GitHub Flavored Markdown) compatibility. It distinguishes itself from siblings like 'markdown-to-html' by being an editor rather than a converter. However, it doesn't explicitly mention what makes it unique among other editing tools in the list.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It doesn't mention prerequisites, target users, or specific scenarios where this editor is preferred over other tools like 'markdown-to-html' or general text editors. The 'Browser-based tool' note is technical but doesn't offer usage context.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
markdown-to-htmlMarkdown-HTML ConverterCInspect
Convert between Markdown and HTML formats.
| Name | Required | Description | Default |
|---|---|---|---|
| input | Yes | Markdown or HTML string to convert | |
| direction | No | Conversion direction | md-to-html |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It states the conversion function but doesn't mention any behavioral traits such as error handling, performance limits, input size constraints, or output format details. For a tool with no annotations, this leaves significant gaps in understanding how it behaves beyond the basic operation.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise and front-loaded: 'Convert between Markdown and HTML formats.' It uses a single, direct sentence with zero waste, making it easy to understand quickly. Every word earns its place by clearly stating the core function.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (format conversion with two parameters) and lack of annotations and output schema, the description is incomplete. It doesn't cover behavioral aspects, usage context, or output details, leaving the agent with insufficient information to fully understand how to invoke and interpret results. For a tool with no structured support beyond the input schema, more descriptive context is needed.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 100% description coverage, with clear documentation for both parameters ('input' and 'direction'), including enum values and defaults. The description adds no additional parameter semantics beyond what the schema provides, such as examples or edge cases. With high schema coverage, the baseline score of 3 is appropriate, as the schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Convert between Markdown and HTML formats.' It specifies the verb ('convert') and resources ('Markdown and HTML formats'), making the function unambiguous. However, it doesn't explicitly differentiate from sibling tools like 'format-converter' or 'xml-json-converter', which handle other format conversions, so it misses full sibling distinction.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It doesn't mention any prerequisites, constraints, or recommend other tools for related tasks (e.g., 'format-converter' for other formats or 'markdown-editor' for editing). Usage is implied by the name and purpose but not explicitly stated.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
math-evaluatorMath EvaluatorCInspect
Evaluate mathematical expressions with variables and functions.
| Name | Required | Description | Default |
|---|---|---|---|
| precision | No | Decimal precision | |
| variables | No | Custom variables (e.g. {"x": 5, "y": 10}) | |
| expression | Yes | Math expression to evaluate (e.g. "2+3*4", "sin(PI/2)", "sqrt(144)") |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions the tool evaluates expressions but doesn't describe what happens with invalid inputs, error handling, performance characteristics, or output format. For a tool with no annotation coverage, this leaves significant gaps in understanding its behavior.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise (one sentence) and front-loaded with the core purpose. Every word earns its place, with no redundant information or fluff. It efficiently communicates the essential function without unnecessary elaboration.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the lack of annotations and output schema, the description is incomplete. It doesn't explain what the tool returns (e.g., numeric result, error message), how it handles edge cases, or any dependencies. For a tool with three parameters and no structured output documentation, more context is needed to understand its full behavior.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all parameters thoroughly. The description adds minimal value beyond the schema by mentioning 'variables and functions' in general terms, but doesn't provide additional syntax, examples, or constraints. This meets the baseline for high schema coverage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's function ('evaluate mathematical expressions') and specifies the types of inputs it handles ('with variables and functions'). It distinguishes itself from most sibling tools by focusing on mathematical evaluation, though it doesn't explicitly differentiate from potential similar tools like 'calculator' or 'unit-converter' that might exist elsewhere.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It doesn't mention any prerequisites, limitations, or suggest other tools for related tasks (e.g., using 'unit-converter' for unit conversions or 'grade-calculator' for academic calculations). The usage context is implied but not explicitly stated.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
mercari-calculatorMercari Profit CalcBInspect
Calculate net profit after 10% fee and shipping costs.
| Name | Required | Description | Default |
|---|---|---|---|
| costPrice | No | Cost/purchase price in JPY | |
| salePrice | Yes | Sale price in JPY | |
| shippingCost | No | Shipping cost in JPY |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It states the calculation logic (10% fee and shipping costs) but doesn't disclose behavioral traits like whether this is a read-only operation, if it requires authentication, rate limits, or what format the output takes. The description is minimal and lacks operational context.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence that states the tool's purpose clearly with zero waste. It's appropriately sized and front-loaded, making it easy to understand at a glance.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's low complexity (simple calculation), 100% schema coverage, and no output schema, the description is minimally adequate. However, with no annotations and no output details, it lacks completeness for operational use, such as explaining the output format or error handling.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, with all parameters well-documented in the schema (costPrice, salePrice, shippingCost in JPY). The description adds no additional parameter semantics beyond implying the calculation uses these inputs, so it meets the baseline of 3 where the schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's function with a specific verb ('calculate') and resource ('net profit'), and specifies the fee structure (10%) and shipping costs. However, it doesn't explicitly differentiate from sibling tools like 'profit-loss' or 'shopify-profit-calculator' that might serve similar purposes.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives. The description doesn't mention any prerequisites, context for use, or comparison with sibling tools like 'mercari-pricing-guide' or 'mercari-shipping-compare' that might be related.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
mercari-listing-templateメルカリ出品テンプレートBInspect
カテゴリ別の商品説明テンプレを一瞬で生成 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It only adds the note '(Browser-based tool)' without explaining what that implies for execution (e.g., does it return HTML, open a browser window, or require user interaction?). It does not disclose side effects, auth requirements, or return value format.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description consists of a single efficient sentence that immediately conveys the core function. There is no redundant information or wasted prose; every token earns its place.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the absence of an output schema and annotations, the description fails to specify what the tool returns (text? HTML? a file?) or what categories are supported. For a template generation tool, omitting the output format significantly hinders the agent's ability to predict how to integrate the result into a workflow.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema contains zero parameters. According to scoring rules, zero parameters establishes a baseline score of 4, as there are no parameter semantics to clarify beyond what the empty schema already communicates.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool generates 'category-specific product description templates' (カテゴリ別の商品説明テンプレを生成), providing a specific verb and resource. However, it relies on the tool name to imply this is for Mercari specifically, and does not explicitly differentiate from siblings like 'ec-template-generator' or 'product-description-generator'.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no explicit guidance on when to use this tool versus alternatives (e.g., when to use this instead of 'mercari-calculator' or general 'product-description-generator'). While 'category-specific' implies usage for categorized items, there are no clear conditions, prerequisites, or alternative recommendations.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
mercari-pricing-guideメルカリ値段設定ガイドBInspect
利益率から最適価格を逆算。送料込み/別の比較 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full disclosure burden. It adds the crucial context that this is a 'Browser-based tool', implying an interactive/GUI experience rather than a data-in/data-out API call. However, it lacks details on what data will be requested interactively, return format, or any side effects.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely efficient: two short clauses delivering the calculation method, feature scope (shipping comparison), and technical modality (browser-based). No filler content.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given zero parameters, no annotations, and no output schema, the description is minimally adequate but thin. For a pricing calculation tool, it should ideally clarify what outputs or guidance the browser interface provides, as the agent cannot infer the tool's utility beyond 'pricing guide' without invocation.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema has zero parameters, establishing a baseline score of 4. The description does not need to describe parameters but hints at why none are present via the 'Browser-based' designation, suggesting user interaction will handle inputs.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the core function (reverse-calculating optimal price from profit margin) and a specific feature (comparing with/without shipping costs). However, it does not explicitly differentiate from sibling tools like 'mercari-calculator' or 'mercari-shipping-compare'.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to select this tool versus the four other Mercari-related siblings (calculator, listing template, sales tracker, shipping compare). No prerequisites or exclusions are mentioned.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
mercari-sales-trackerメルカリ売上管理表BInspect
月別売上/利益/手数料をlocalStorage管理。確定申告対応CSV出力 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations, description carries the burden. It successfully discloses localStorage (client-side persistence) and browser-based execution, but omits details on UI behavior, data lifecycle, or what occurs upon invocation.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely compact two-phrase structure with zero filler. Front-loaded with core functionality (sales/fee management) and parenthetical clarifies execution environment without clutter.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Sufficient for a parameterless tool—explains the core tracking and export functionality. However, given no output schema and no annotations, it should clarify what the tool returns (HTML UI? data?) and how it differs from mercari-calculator.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Zero parameters present; per scoring rules, this establishes a baseline score of 4 with no deductions needed.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Clearly identifies the resource (monthly sales/profit/fees) and action (localStorage management), distinguishing from calculation-focused siblings like mercari-calculator by emphasizing persistence (localStorage) and tax reporting output.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Provides a hint at usage context (確定申告対応/tax return compatible CSV), but lacks explicit when-to-use guidance versus siblings or prerequisites like data entry requirements.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
mercari-shipping-compareメルカリ発送方法比較AInspect
サイズ・重さ入力→最安発送方法を自動判定。全配送方法比較 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description must carry the full behavioral burden. It successfully discloses the 'Browser-based tool' nature, indicating interactive behavior where inputs are likely collected via browser rather than parameters. However, it omits other behavioral details like whether it requires Mercari authentication, network access, or what data persists.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is exceptionally concise with two efficient Japanese phrases and a parenthetical note. Every element earns its place: the input→output workflow, the scope (all methods), and the execution mode (browser-based). No redundancy or filler.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the empty parameter schema, lack of annotations, and no output schema, the description covers the essential functionality but remains minimal. It adequately explains what the tool does but omits details about return format, error conditions, or specific Mercari shipping carriers included in the comparison.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema has 0 parameters, establishing a baseline of 4. The description adds value by clarifying that while there are no MCP parameters, the tool expects size and weight inputs to be provided interactively in the browser context, resolving the apparent mismatch between the description and empty schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool compares Mercari shipping methods and automatically determines the cheapest option based on size/weight inputs. It identifies the specific resource (shipping methods) and action (comparison/automatic determination), distinguishing it from the generic 'shipping-calculator' sibling by being Mercari-specific. However, it could better clarify how it differs from 'mercari-calculator'.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage through the workflow 'size/weight input → cheapest method determination', suggesting when to use it (when you have package dimensions). However, it lacks explicit guidance on when NOT to use it or when to prefer siblings like 'shipping-calculator' or 'mercari-pricing-guide'.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
meta-description-generatorメタディスクリプション生成BInspect
SEOに効くdescriptionを120/160文字で自動生成 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
The description adds 'Browser-based tool' and specifies character constraints (120/160), which supplements the missing annotations. However, it fails to clarify how the tool operates with zero input parameters (what source content it processes) or what the return format is.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is efficiently structured in a single sentence with the core value proposition ('SEOに効くdescription') front-loaded. The parenthetical '(Browser-based tool)' adds distinct behavioral info without excessive length.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a zero-parameter tool without output schema or annotations, the description adequately covers the generation purpose and constraints. However, it lacks critical behavioral details regarding data sources (since no inputs are defined) and return values, leaving operational ambiguity.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema contains zero parameters. According to the rubric, zero-parameter tools receive a baseline score of 4. The description does not need to compensate for missing parameter documentation in this case.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool generates 'SEO-effective descriptions' at '120/160 characters', providing specific scope and resource details. However, it does not explicitly distinguish from sibling text-generation tools (e.g., article-outline-generator) beyond the SEO context.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to select this tool versus alternatives like 'og-preview' or other content generators. It lacks prerequisites, input requirements (notably despite having zero schema parameters), or workflow context.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
nda-generatorNDA GeneratorBInspect
Generate NDAs based on Japan METI templates in 30 seconds. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It mentions 'in 30 seconds' (suggesting speed) and 'Browser-based tool' (implying client-side execution), but it lacks critical behavioral details: what the output format is (e.g., PDF, Word), whether it requires authentication, if it's read-only or mutative, or any rate limits. For a tool with zero annotation coverage, this leaves significant gaps.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise and front-loaded: it states the core functionality in the first clause, adds context (template source and speed), and includes an environmental note in parentheses. Every sentence earns its place with no wasted words, making it easy to scan and understand quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool has 0 parameters and no output schema, the description is minimally adequate. It covers the purpose and some behavioral hints (speed, browser-based), but lacks details on output format, error handling, or integration context. Without annotations, it should provide more behavioral transparency to be fully complete for a generation tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description adds value by specifying the template source ('Japan METI templates') and time frame ('30 seconds'), which are not captured in the schema. However, it doesn't detail any optional inputs or customization options, though none are required.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Generate NDAs based on Japan METI templates in 30 seconds.' It specifies the verb ('Generate'), resource ('NDAs'), and scope ('based on Japan METI templates'), distinguishing it from generic document generators. However, it doesn't explicitly differentiate from sibling tools like 'contract-generator' or 'terms-generator' beyond mentioning the specific template source.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides minimal usage guidance. It implies this tool is for creating NDAs using Japanese METI templates quickly, but it doesn't specify when to use it versus alternatives (e.g., 'contract-generator'), prerequisites, or exclusions. The mention 'Browser-based tool' hints at the execution environment but doesn't offer operational context.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
nhi-calc国民健康保険シミュレーターBInspect
主要10都市対応。自治体間の保険料比較も一瞬 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full disclosure burden. It notes the tool is 'Browser-based' but fails to explain critical behavioral aspects: what the tool returns (URL? HTML content?), whether user interaction is required, data persistence, or authentication requirements. The zero-parameter schema makes behavioral expectations especially opaque.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Description is extremely concise with two distinct sentences. Neither sentence is redundant or tautological. However, given the heavy lifting required (no annotations, no output schema, zero parameters), this level of brevity leaves critical gaps, preventing a score of 5.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the absence of annotations, output schema, and input parameters, the description inadequately explains how a parameterless insurance calculator functions. It omits return format, execution flow, and prerequisite conditions necessary for agent invocation without structured metadata support.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema contains zero parameters. Per scoring guidelines, zero-parameter tools receive a baseline score of 4. The description appropriately does not invent parameter semantics where none exist.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description clearly identifies the tool as a National Health Insurance (NHI) premium calculator supporting 10 major cities and municipality comparisons. However, it lacks explicit distinction from siblings like 'social-insurance-calc' (社会保険) or 'resident-tax-calc', which handle related but distinct financial domains.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance provided on when to select this tool versus alternatives. Given siblings include multiple tax and insurance calculators (social-insurance-calc, freelance-tax-calc, resident-tax-calc), the description fails to specify target users (e.g., self-employed individuals vs. employees).
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
npv-irr-calcNPV/IRR CalculatorCInspect
Net present value and internal rate of return calculation.
| Name | Required | Description | Default |
|---|---|---|---|
| cashFlows | Yes | Cash flow entries (year 0 should be negative for initial investment) | |
| discountRate | No | Discount rate (%) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden for behavioral disclosure. It states what the tool calculates but doesn't describe output format, error conditions, computational limitations, or whether it's a read-only vs. write operation. For a financial calculation tool with no annotation coverage, this leaves significant behavioral gaps.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise - a single sentence that states the core function without unnecessary words. It's front-loaded with the essential information and wastes no space on redundant details.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a financial calculation tool with no annotations and no output schema, the description is insufficient. It doesn't explain what the outputs look like (NPV value, IRR percentage), how results are formatted, or any assumptions about the calculations. The context signals show this is a non-trivial tool with array parameters, yet the description provides minimal context.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The schema has 100% description coverage, thoroughly documenting both parameters (cashFlows array with year/amount details and discountRate percentage). The description adds no parameter information beyond what's in the schema, so it meets the baseline for high schema coverage but doesn't provide additional semantic context.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description states the tool calculates net present value and internal rate of return, which is a clear purpose. However, it doesn't specify what resources or inputs are involved beyond the name, and it doesn't distinguish this from sibling tools like 'investment-simulator' or 'bond-yield-calc' that might perform similar financial calculations.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It doesn't mention prerequisites, typical scenarios, or compare it to sibling tools like 'investment-simulator' or 'cash-flow-statement' that might overlap in financial analysis contexts.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
og-previewOG Image PreviewBInspect
Preview how links appear when shared on social media. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions 'Browser-based tool,' which implies it might open a browser or simulate one, but doesn't specify whether it's read-only, requires internet access, has rate limits, or what the output looks like (e.g., a screenshot or metadata). For a tool with zero annotation coverage, this leaves significant gaps in understanding its behavior.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise and front-loaded: the first sentence states the core purpose, and the parenthetical adds implementation context. Every word earns its place with zero waste, making it easy for an AI agent to parse quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool has 0 parameters, no annotations, and no output schema, the description is minimally adequate. It explains what the tool does and adds 'Browser-based tool' for context, but lacks details on behavioral traits (e.g., output format, side effects) that would be helpful for an agent. It meets the bare minimum for a zero-parameter tool but doesn't fully compensate for the lack of structured data.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description doesn't add parameter details, which is appropriate. A baseline of 4 is given because the schema fully covers the absence of parameters, and the description doesn't need to compensate for any gaps.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Preview how links appear when shared on social media.' It specifies the verb ('preview') and resource ('links'), and the parenthetical 'Browser-based tool' adds useful context about the implementation. However, it doesn't explicitly differentiate from sibling tools like 'meta-description-generator' or 'thumbnail-ctr-predictor' that might also relate to social media content.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It doesn't mention prerequisites (e.g., needing a URL), exclusions, or comparisons to sibling tools like 'meta-description-generator' or 'sns-post-template' that might overlap in social media context. The 'Browser-based tool' note hints at technical constraints but doesn't constitute usage guidance.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
panorama-splitterPanorama SplitterAInspect
Split wide images into Instagram carousel slides. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions 'Browser-based tool,' which hints at execution context, but does not describe how the splitting works (e.g., automatic vs. manual, aspect ratio handling, output format), potential side effects (e.g., image quality loss), or any prerequisites (e.g., supported image types). This leaves significant gaps for a tool that modifies images.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise—two short phrases that directly state the tool's purpose and execution context. Every word earns its place, with no redundancy or fluff, making it easy to understand at a glance.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (image processing with no parameters) and lack of annotations or output schema, the description is minimal but functional. It covers the basic purpose and context but lacks details on behavior, output format, or limitations, which would be helpful for an agent to use it effectively. It is complete enough for a simple tool but could be more informative.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description does not add parameter details, which is appropriate, but it could have mentioned implicit inputs (e.g., image uploads via browser). Since there are no parameters, a baseline of 4 is applied, as the description adequately addresses the tool's function without unnecessary parameter elaboration.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the specific action ('Split wide images') and the target resource ('Instagram carousel slides'), with the parenthetical '(Browser-based tool)' providing additional context about the execution environment. It effectively distinguishes itself from sibling tools like 'image-crop' or 'image-resizer' by focusing on a specialized use case for social media formatting.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage for converting wide images into Instagram-optimized slides, but does not explicitly state when to use this tool versus alternatives (e.g., 'image-crop' for general cropping or 'smart-resize' for resizing). It provides a clear context (Instagram carousel creation) but lacks explicit exclusions or comparisons to sibling tools.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
password-generatorパスワード生成BInspect
安全なパスワードを即座に生成。長さ・文字種をカスタマイズ
| Name | Required | Description | Default |
|---|---|---|---|
| count | No | Number of passwords to generate | |
| length | No | Password length | |
| numbers | No | Include numbers | |
| symbols | No | Include symbols | |
| lowercase | No | Include lowercase letters | |
| uppercase | No | Include uppercase letters |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It mentions '安全なパスワード' (secure passwords) and '即座に生成' (instant generation), but doesn't specify what makes passwords secure, whether they're cryptographically random, if there are rate limits, or what the output format looks like. For a tool with zero annotation coverage, this leaves significant behavioral gaps.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise with just two short phrases that efficiently communicate the core functionality. Every word earns its place, and the information is front-loaded with the primary purpose stated first. No wasted words or unnecessary elaboration.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a relatively simple tool with 100% schema coverage but no annotations and no output schema, the description provides the minimum viable information. It states what the tool does but lacks important context about security guarantees, output format, and usage constraints. The description is adequate but has clear gaps in behavioral transparency.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already fully documents all 6 parameters with clear descriptions. The description adds minimal value beyond the schema by mentioning '長さ・文字種をカスタマイズ' (customize length and character types), which aligns with the 'length', 'numbers', 'symbols', 'lowercase', and 'uppercase' parameters but doesn't provide additional semantic context.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: '安全なパスワードを即座に生成' (generate secure passwords instantly) and specifies customization capabilities ('長さ・文字種をカスタマイズ' - customize length and character types). It distinguishes itself from sibling tools by focusing specifically on password generation, though it doesn't explicitly differentiate from similar utilities like 'uuid-generator' or 'hash-generator'.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided about when to use this tool versus alternatives. The description doesn't mention any prerequisites, constraints, or suggest when other tools might be more appropriate. It simply states what the tool does without contextual usage information.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
payslip-generatorPayslip GeneratorBInspect
Payslips with auto social insurance calculation. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden for behavioral disclosure. It mentions 'Browser-based tool' which provides useful implementation context, but doesn't disclose critical behavioral traits: whether it generates downloadable files, requires authentication, has rate limits, or what format the output takes. For a document generation tool with zero annotation coverage, this represents significant gaps in behavioral transparency.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise at just 8 words, with zero wasted language. It's front-loaded with the core functionality ('Payslips with auto social insurance calculation') followed by implementation context. While efficient, it borders on being too terse given the lack of behavioral details needed for a tool with no annotations.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given that this is a document generation tool with no annotations, no output schema, and zero parameters, the description is incomplete. While it states what the tool does, it doesn't explain what happens when invoked: does it open a browser interface? Generate a PDF? Return HTML? For a tool with no structured metadata, the description should provide more complete context about the expected user experience and output.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The tool has zero parameters with 100% schema description coverage, so the baseline is 4. The description doesn't need to explain parameters, and it doesn't attempt to. The mention of 'auto social insurance calculation' could be seen as hinting at implicit inputs, but since the schema explicitly shows no parameters, this doesn't detract from the score.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: generating payslips with automated social insurance calculation. It specifies the verb ('generate') and resource ('payslips'), and distinguishes itself from siblings like 'take-home-pay-calc' or 'salary-vs-freelance' by focusing on payslip generation rather than pure calculation. However, it doesn't explicitly differentiate from 'invoice-generator' or 'receipt-generator', which might share similar document-generation functions.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides minimal guidance on when to use this tool. It mentions 'auto social insurance calculation' which implies it's for payroll contexts, but doesn't specify when to choose it over alternatives like 'take-home-pay-calc' or 'social-insurance-calc'. No explicit when-not-to-use scenarios or prerequisites are mentioned, leaving usage context largely implied rather than clearly defined.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
pdf-compressPDF CompressBInspect
Reduce PDF file size for email attachments. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions 'Browser-based tool,' which hints at client-side processing, but fails to detail critical aspects like whether the tool modifies the original file, handles large files, supports specific PDF versions, or has rate limits. For a tool with zero annotation coverage, this is insufficient.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise and front-loaded: 'Reduce PDF file size for email attachments. (Browser-based tool).' Both sentences are relevant and add value without waste, making it easy to scan and understand quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's simplicity (0 parameters, no output schema, no annotations), the description is adequate but not fully complete. It covers the basic purpose and hints at the environment ('Browser-based'), but lacks details on behavioral traits, output format, or error handling. For a tool with no structured data, it meets minimum viability but has clear gaps.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description does not add parameter details, which is appropriate. A baseline of 4 is applied because the schema fully covers the parameters (none), and the description does not introduce confusion or redundancy.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Reduce PDF file size for email attachments.' It specifies the verb ('Reduce'), resource ('PDF file size'), and context ('for email attachments'), making it easy to understand. However, it does not explicitly differentiate from sibling tools like 'pdf-join' or 'pdf-split', which prevents a score of 5.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides minimal guidance by mentioning 'for email attachments,' which implies a use case, but it lacks explicit instructions on when to use this tool versus alternatives (e.g., other PDF tools in the sibling list). There is no mention of prerequisites, limitations, or comparisons, leaving gaps in usage context.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
pdf-joinPDF MergeCInspect
Combine multiple PDFs into one with drag-to-reorder.
| Name | Required | Description | Default |
|---|---|---|---|
| pdfs | Yes | Array of base64-encoded PDF files to merge (order preserved) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It mentions 'drag-to-reorder' which implies interactive reordering, but doesn't disclose critical behavioral traits: whether this is a read-only or destructive operation, authentication needs, rate limits, output format, or error handling. For a tool that merges files (potentially destructive), this is inadequate.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise (one sentence) and front-loaded with the core purpose. Every word earns its place: 'Combine multiple PDFs into one' states the action, and 'with drag-to-reorder' adds valuable context without redundancy. No wasted words or unnecessary elaboration.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given this is a file manipulation tool with no annotations and no output schema, the description is incomplete. It doesn't explain what the merged output looks like (e.g., base64 string, file download), error conditions, or behavioral constraints. For a tool that could have significant side effects (merging files), more context is needed.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, with the single parameter 'pdfs' well-documented in the schema as 'Array of base64-encoded PDF files to merge (order preserved)'. The description adds 'drag-to-reorder' which implies reordering capability, but doesn't provide additional parameter semantics beyond what the schema already covers. Baseline 3 is appropriate when schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Combine multiple PDFs into one' specifies the verb (combine/merge) and resource (PDFs). It distinguishes from sibling tools like pdf-split or pdf-compress by focusing on merging. However, it doesn't explicitly differentiate from all PDF-related siblings beyond the title, so it's not a perfect 5.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides minimal usage guidance. 'with drag-to-reorder' hints at reordering capability, but doesn't specify when to use this tool versus alternatives like pdf-stamper or pdf-redact. No explicit when/when-not instructions or prerequisites are mentioned, leaving the agent with little contextual guidance.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
pdf-metadata-removerPDF Metadata RemoverAInspect
Strip author, edit history, and metadata before sharing. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden. It discloses that the tool strips metadata (a destructive action) and is browser-based, which adds useful context about its operational environment. However, it doesn't specify what happens to the processed PDF (e.g., if it's downloaded or modified in-place), any limitations (e.g., file size), or error handling, leaving gaps in behavioral understanding.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise—just one sentence with two clauses—and front-loaded with the core action ('Strip author, edit history, and metadata'). Every word earns its place, providing essential information without waste.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (a destructive operation with no parameters), no annotations, and no output schema, the description is somewhat complete but lacks details. It covers the what and why (metadata removal for sharing) and the how (browser-based), but doesn't explain the output (e.g., what the user receives after processing) or potential side effects, leaving room for improvement.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The tool has 0 parameters with 100% schema description coverage, so no parameter documentation is needed. The description doesn't add param info, but since there are no parameters, this is acceptable. A baseline of 4 is appropriate as it doesn't need to compensate for any schema gaps.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with specific verbs ('Strip') and resources ('author, edit history, and metadata'), and identifies the target context ('before sharing'). It doesn't explicitly differentiate from sibling tools like 'pdf-redact' or 'pdf-password', but the focus on metadata removal is distinct enough for general clarity.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage context ('before sharing') and mentions it's 'Browser-based', which suggests it operates locally or in a browser environment. However, it lacks explicit guidance on when to use this versus alternatives like 'pdf-redact' or 'exif-remover', and no exclusions or prerequisites are stated.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
pdf-passwordPDF PasswordBInspect
Encrypt PDFs with password protection. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It states the tool encrypts PDFs with passwords, implying a write/mutation operation, but doesn't disclose behavioral traits like whether it requires specific inputs (e.g., file uploads), what happens to the original file, error handling, or rate limits. The 'Browser-based tool' note adds some context about execution but is insufficient for a mutation tool.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise—two short phrases that directly state the core function and an additional note. Every word earns its place with no redundancy or fluff, making it easy to parse quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (encryption implies mutation), lack of annotations, and no output schema, the description is incomplete. It doesn't explain how the tool works (e.g., input method, output format), security implications, or error cases. For a mutation tool with zero structured data support, this leaves significant gaps for an AI agent.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameters need documentation. The description doesn't add param info, which is fine here. Baseline is 4 for 0 parameters, as it avoids confusion by not discussing non-existent params.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with a specific verb ('Encrypt') and resource ('PDFs'), plus the additional function 'with password protection'. It distinguishes from siblings like pdf-compress or pdf-split by focusing on encryption. However, it doesn't explicitly differentiate from pdf-unlock (which might handle password removal), keeping it from a perfect 5.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It mentions 'Browser-based tool', which hints at the execution environment but doesn't specify prerequisites, limitations, or when to choose it over other PDF tools. No explicit when/when-not or alternatives are stated.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
pdf-redactPDF RedactAInspect
Black out sensitive text in PDFs without server processing. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden. It discloses key behavioral traits: it performs a destructive action ('Black out') implying permanent text removal, and specifies 'without server processing' indicating local execution for privacy. However, it misses details like supported PDF formats, whether it preserves other content, or error handling, leaving gaps for a mutation tool.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise with two brief clauses, front-loading the core purpose ('Black out sensitive text in PDFs') and adding a useful qualifier. Every word earns its place, with no redundancy or fluff, making it highly efficient and well-structured for quick understanding.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (a mutation operation with no annotations and no output schema), the description is moderately complete. It covers the what and how (browser-based), but lacks details on behavioral aspects like reversibility, supported PDF versions, or output format. For a redaction tool, more context on limitations or results would improve completeness.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description adds no param info, which is appropriate. Baseline is 4 for 0 params, as it doesn't need to compensate for schema gaps, but it doesn't reach 5 since it doesn't explain how the tool operates without inputs (e.g., via UI interaction).
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the action ('Black out sensitive text') and resource ('PDFs'), specifying it's for redaction. It distinguishes from siblings like 'pdf-compress' or 'pdf-join' by focusing on content modification rather than file operations. However, it doesn't explicitly differentiate from all PDF tools (e.g., 'pdf-sign' also modifies PDFs), keeping it from a perfect 5.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage for privacy/security needs by mentioning 'sensitive text' and notes it's 'browser-based', suggesting local processing vs. server tools. However, it lacks explicit when-not-to-use guidance or named alternatives among siblings (e.g., vs. 'pdf-metadata-remover' for different privacy aspects), leaving usage context somewhat implied rather than fully articulated.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
pdf-rotatePDF RotateBInspect
Rotate PDF pages by 90/180/270 degrees. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions 'Browser-based tool', which adds some context about the environment, but fails to disclose critical traits like whether the rotation is destructive to the original file, authentication needs, rate limits, or output format. This leaves significant gaps for a tool that modifies PDFs.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise—two short phrases that directly state the tool's purpose and context. Every word earns its place, with no wasted sentences, making it front-loaded and efficient for quick understanding.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity of a PDF manipulation tool with no annotations and no output schema, the description is incomplete. It lacks details on behavioral traits (e.g., file handling, errors), output format, and usage context, making it inadequate for safe and effective tool invocation by an AI agent.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameters need documentation. The description adds value by specifying rotation angles (90/180/270 degrees), which clarifies the tool's functionality beyond the empty schema. However, it does not explain how to select or apply these angles, leaving minor ambiguity.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the action ('Rotate PDF pages') and specifies the rotation angles (90/180/270 degrees), which is a specific verb+resource combination. However, it does not explicitly differentiate from sibling tools like 'pdf-split' or 'pdf-join', though the function is distinct by nature. It avoids tautology by not merely restating the name/title.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives, such as other PDF manipulation tools in the sibling list (e.g., 'pdf-split', 'pdf-join'). It mentions it's 'Browser-based', which hints at the context but does not specify prerequisites, limitations, or explicit alternatives, leaving usage unclear.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
pdf-signPDF SignBInspect
Add handwritten or text signatures directly to PDFs. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions the tool is 'Browser-based', which adds some context about the execution environment, but it fails to describe key traits like whether it modifies the original PDF, requires user interaction for signatures, handles errors, or has any rate limits. For a tool with zero annotation coverage, this is insufficient.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise: one main sentence stating the purpose and a brief parenthetical note. Every word earns its place, with no redundancy or fluff, making it easy to parse and front-loaded with essential information.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (simple PDF editing), no annotations, no output schema, and 0 parameters, the description is minimally adequate. It covers the basic purpose and environment but lacks details on behavior, output, or error handling. This meets a baseline for a zero-param tool but could be more informative for safe invocation.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description doesn't add param info, which is appropriate. A baseline of 4 is given since the schema fully handles parameters, and the description doesn't need to compensate.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Add handwritten or text signatures directly to PDFs.' It specifies the verb ('Add'), resource ('signatures'), and target ('PDFs'), making the action explicit. However, it doesn't differentiate from sibling tools like 'pdf-stamper' or 'pdf-redact', which might have overlapping functions, so it doesn't reach a perfect score.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides minimal guidance with the note '(Browser-based tool)', implying it requires a browser context, but it lacks explicit when-to-use instructions, alternatives (e.g., vs. 'pdf-stamper'), or prerequisites. This leaves the agent with little direction on optimal usage scenarios.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
pdf-splitPDF SplitBInspect
Split PDFs by page range or into individual pages.
| Name | Required | Description | Default |
|---|---|---|---|
| Yes | Base64-encoded PDF file | ||
| splitMode | No | Split mode: each=per page, range=specific pages, even-odd=split by parity | each |
| pageRanges | No | Page ranges for range mode (e.g. "1-3, 5, 7-10"). 1-indexed. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It states the tool's function but lacks behavioral details such as whether it's a read-only or destructive operation, error handling, output format (e.g., returns multiple files), or performance constraints. This is inadequate for a tool that processes files.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence with zero waste. It front-loads the core purpose and is appropriately sized for the tool's complexity.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the lack of annotations and output schema, the description is incomplete. It doesn't explain what the tool returns (e.g., a list of split PDFs, error messages), behavioral traits, or usage context, which are crucial for an AI agent to invoke it correctly in a file-processing scenario.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema fully documents all parameters. The description adds minimal value by hinting at split modes ('by page range or into individual pages'), but doesn't provide additional syntax or usage examples beyond what the schema specifies. Baseline 3 is appropriate.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the specific action ('Split') and resource ('PDFs'), and distinguishes between different split methods ('by page range or into individual pages'). This is precise and distinguishes it from sibling PDF tools like pdf-join or pdf-compress.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives. It doesn't mention prerequisites (e.g., needing a valid PDF), exclusions, or comparisons with other PDF tools in the sibling list, leaving the agent to infer usage context.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
pdf-stamperDigital StampBInspect
Apply Japanese hanko stamps to PDF documents. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It states the tool applies stamps to PDFs and is browser-based, but lacks critical behavioral details: whether it modifies the original PDF, requires uploads, supports batch processing, has file size limits, or outputs a new file. For a tool with zero annotation coverage, this is insufficient transparency.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is two short, efficient sentences. The first sentence states the core purpose, and the second adds essential context (browser-based). Every word earns its place with no redundancy or fluff.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no annotations, no output schema, and a tool that likely modifies PDFs (implied by 'Apply'), the description is incomplete. It lacks information on behavioral traits (e.g., destructive vs. non-destructive), output format, error handling, or limitations. For a tool with potential side effects, this leaves significant gaps.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description doesn't add parameter details, which is appropriate. Baseline is 4 for zero parameters, as the schema fully covers the absence of inputs.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the action ('Apply Japanese hanko stamps') and resource ('to PDF documents'), with a specific cultural context (Japanese hanko). It distinguishes from siblings like 'pdf-sign' (likely digital signatures) and 'stamp-maker' (likely creating stamps). However, it doesn't explicitly differentiate from all PDF manipulation tools, keeping it at a 4 rather than a 5.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives like 'pdf-sign' or 'stamp-maker'. It mentions it's 'Browser-based', which hints at execution environment but doesn't clarify use cases, prerequisites, or exclusions. This leaves the agent with minimal contextual direction.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
pdf-to-imagePDF to ImageBInspect
Convert PDF pages to high-quality PNG, JPEG, or WebP. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It mentions 'high-quality' and 'browser-based,' hinting at output quality and execution environment, but lacks details on permissions, rate limits, error handling, or what 'browser-based' entails operationally, which is insufficient for a tool with potential resource implications.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise—one sentence with no wasted words. It front-loads the core functionality and includes a brief qualifier, making it easy to parse quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no annotations, 0 parameters, and no output schema, the description is minimally adequate. It covers the basic purpose but lacks behavioral details (e.g., how conversion works, output specifics) that would help an agent use it effectively in varied contexts.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description doesn't add param info, but that's acceptable here, earning a baseline score above 3 due to the lack of parameters.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: converting PDF pages to image formats (PNG, JPEG, WebP). It specifies the resource (PDF pages) and output formats, though it doesn't explicitly differentiate from siblings like 'image-to-pdf' or 'pdf-join' beyond format conversion scope.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives. The description mentions it's 'browser-based,' but doesn't explain implications or compare it to other PDF/image tools in the sibling list, leaving usage context unclear.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
pdf-unlockPDF UnlockAInspect
Remove edit/print/copy restrictions from PDFs. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden for behavioral disclosure. While it states the core function (removing restrictions) and execution context (browser-based), it lacks important details: whether this modifies the original file or creates a new one, what permissions are required, whether it handles encrypted PDFs, what happens if restrictions can't be removed, or what the output looks like. For a tool that alters document security properties, this represents significant gaps in behavioral transparency.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise (two short phrases) with zero wasted words. The main function is stated first, followed by execution context. Every element earns its place, and the structure is front-loaded with the primary purpose.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a zero-parameter tool with no output schema, the description provides the core purpose and execution context. However, as a tool that modifies document security properties, it should ideally include more behavioral details (what happens to the original file, error conditions, output format) given the absence of annotations. The description is minimally adequate but leaves important questions unanswered for a security-related operation.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
With 0 parameters and 100% schema description coverage, the baseline would be 4 even with no parameter information in the description. The description doesn't discuss parameters (since there are none), which is appropriate. The mention of 'Browser-based tool' provides useful context about how the tool operates, though this isn't strictly parameter semantics.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the specific action ('Remove edit/print/copy restrictions') and resource ('from PDFs'), with the parenthetical '(Browser-based tool)' providing additional context about the execution environment. It effectively distinguishes this tool from sibling PDF tools like pdf-password, pdf-redact, or pdf-compress by focusing on restriction removal rather than other PDF operations.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage context (when PDFs have restrictions that need removal) but doesn't explicitly state when to use this tool versus alternatives like pdf-password (which might add/remove password protection) or other PDF manipulation tools. No explicit exclusions or prerequisites are mentioned, leaving some ambiguity about appropriate use cases.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
pharma-law-checker薬機法・景表法チェッカーBInspect
NGワード1000語+をローカル照合。原稿を外部に送らない (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations, the description carries the full burden. It successfully discloses critical privacy behavior: '原稿を外部に送らない' (manuscript not sent externally) and 'ローカル照合' (local matching). However, it fails to describe output behavior—whether it returns violation lists, risk scores, or highlighted text—which is essential for a compliance tool with no output schema.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely concise with zero wasted words. The single sentence packs specific technical details (1000+ word database, local processing, privacy guarantee) and the parenthetical adds execution context without bloating the description. Every element earns its place.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given zero parameters and no output schema, the description covers the unique value proposition (privacy/local processing) but leaves significant gaps. For a specialized compliance tool, it should describe the return format (e.g., 'returns flagged violations with risk levels') and explain the zero-parameter invocation pattern. The privacy disclosure prevents a lower score.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema contains zero parameters, establishing a baseline score of 4. The description does not attempt to explain non-existent parameters, which is appropriate. However, it does not clarify how text input is handled given the empty parameter list (e.g., interactive mode, clipboard access), which remains ambiguous.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly identifies the tool's function: checking against pharmaceutical law (薬機法) and prize display act (景表法) using a local database of 1000+ prohibited words (NGワード). The mention of '原稿' (manuscript) clarifies the target resource. However, the 'Browser-based tool' parenthetical slightly confuses the execution context without explaining how it relates to MCP invocation.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance provided on when to use this versus general text checkers or compliance tools. Given the specialized legal domain (pharmaceutical advertising), explicit guidance like 'use when reviewing medical/pharmaceutical advertising copy for compliance' would be expected but is absent.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
photo-ledger-maker写真台帳メーカーBInspect
写真をドロップして台帳PDFを自動生成 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It mentions 'Browser-based tool' which adds some context about the execution environment, but fails to disclose critical behavioral traits: whether it modifies/overwrites files, requires specific permissions, handles errors, or produces downloadable output. For a tool with zero annotation coverage, this is a significant gap.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise: two parts (Japanese action statement + English context note) in one sentence. It's front-loaded with the core functionality and wastes no words. Every element ('写真をドロップして台帳PDFを自動生成', 'Browser-based tool') adds value.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given 0 parameters and no output schema, the description covers the basic purpose adequately. However, as a tool with no annotations, it should provide more behavioral context (e.g., what happens to uploaded photos, output format details). The 'Browser-based tool' note helps but doesn't fully compensate for missing safety/operational details.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description appropriately doesn't discuss parameters, focusing instead on the tool's high-level functionality. Baseline 4 is applied as it avoids unnecessary parameter details.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: '写真をドロップして台帳PDFを自動生成' (Drop photos to automatically generate a ledger PDF). It specifies the action (drop photos) and outcome (generate PDF ledger), though it doesn't explicitly differentiate from sibling tools like 'image-to-pdf' or 'pdf-join' beyond mentioning 'ledger' format.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides minimal guidance: it mentions 'Browser-based tool' which hints at usage context, but offers no explicit when-to-use rules, alternatives, or exclusions compared to similar tools (e.g., 'image-to-pdf', 'pdf-join'). It lacks prerequisites or scenarios for optimal use.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
product-description-generator商品説明文ジェネレーターBInspect
商品スペック入力→プラットフォーム別の出品用説明文を即生成 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full disclosure burden. It adds the critical behavioral context that this is a 'Browser-based tool', signaling that interaction occurs via browser rather than headless API. However, it omits which e-commerce platforms are supported, whether descriptions are returned as text or saved to a file, and any data retention policies.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The single-line description uses efficient arrow notation (→) to denote the input-output flow, keeping it front-loaded with the core value proposition. The '(Browser-based tool)' parenthetical is appropriately placed as metadata. While appropriately concise, it lacks substructure that could have enumerated supported platforms.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complex functionality (AI generation for multiple platforms) and absence of both output schema and annotations, the description is minimally adequate but incomplete. It omits the list of supported e-commerce platforms, the output format (JSON string? file path? inline text?), and the nature of the browser interaction (opens new tab? embedded view?). For a content generation tool with zero MCP parameters, more behavioral context was needed.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema has 0 parameters, establishing a baseline of 4 per the scoring rules. The description clarifies that input occurs ('商品スペック入力') despite the empty schema, which semantically aligns with the disclosed 'Browser-based' nature (inputs captured in UI rather than MCP params). This resolves what would otherwise be ambiguity between the empty schema and the described functionality.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the core function using the flow '商品スペック入力→プラットフォーム別の出品用説明文を即生成' (input specs → generate platform-specific listing descriptions), identifying the verb (generate), resource (listing descriptions), and scope (platform-specific). It sufficiently distinguishes from siblings like 'meta-description-generator' by specifying '出品用' (for product listings), though it lacks explicit differentiation from 'ec-template-generator'.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives like 'ec-template-generator', 'sales-writing-analyzer', or 'rewrite-assistant'. There is no mention of prerequisites (e.g., needing product specs ready) or exclusion criteria (e.g., when text generation might fail).
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
product-photo-studio商品画像スタジオBInspect
白背景+影+正方形トリミングを一括加工。出品画像を一瞬で整える (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Without annotations, the description carries the full burden. It successfully discloses that this is a 'Browser-based tool' (indicating interactive UI rather than headless API) and mentions 'batch processing' (一括加工). However, it omits critical behavioral details: whether it saves files locally, returns image data, opens a new tab, or requires user interaction to complete the process.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is efficiently structured across two sentences with the primary functionality (white bg/shadow/crop) front-loaded in the first sentence. The parenthetical '(Browser-based tool)' is slightly detached but provides essential behavioral context without excessive verbosity.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
With no output schema provided, the description should ideally explain the result of invocation (e.g., downloaded file, browser window opened, image URL returned). It describes the transformation but not the delivery mechanism or persistence model, leaving gaps in the agent's understanding of the complete workflow.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema contains zero parameters. According to the rubric, zero-parameter tools receive a baseline score of 4, as there are no parameters requiring semantic elaboration beyond what the empty schema conveys.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly specifies the exact image processing operations performed (white background, shadow, square cropping) and the target use case (listing images/出品画像). However, it does not explicitly differentiate why one should use this combined tool over chaining sibling tools like bg-remover + image-crop + smart-resize.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to prefer this tool versus the numerous sibling image tools (bg-remover, pro-matting, image-crop, etc.). The phrase 'Browser-based tool' implies an interactive workflow, but does not clarify prerequisites or when this is superior to automated API-based alternatives.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
profit-lossProfit & LossBInspect
Auto-generate P&L statements from revenue and expense data. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions 'auto-generate' but does not specify how the generation works (e.g., automated calculations, template-based), what inputs are needed beyond the implied data, or any constraints like data format requirements. The '(Browser-based tool)' note hints at a user interface but lacks detail on functionality or limitations.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is brief and front-loaded with the core function, followed by a parenthetical note. Both sentences are relevant, with no wasted words. However, the '(Browser-based tool)' note could be more informative, slightly reducing efficiency. Overall, it is well-structured and concise.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (financial statement generation) and lack of annotations and output schema, the description is incomplete. It does not explain how the P&L is generated, what the output looks like, or any behavioral traits like error handling or data requirements. For a tool with no structured support, more detail is needed to guide effective use.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The tool has 0 parameters with 100% schema description coverage, so no parameter documentation is needed. The description implies input data ('revenue and expense data') but does not specify parameters, which is acceptable since the schema fully covers the absence of parameters. A baseline of 4 is appropriate as the description does not need to compensate for any parameter gaps.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's function: 'Auto-generate P&L statements from revenue and expense data.' It specifies the verb ('auto-generate'), resource ('P&L statements'), and source data ('revenue and expense data'). However, it does not explicitly distinguish this tool from sibling tools like 'balance-sheet' or 'cash-flow-statement' that also generate financial statements, which prevents a score of 5.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It lacks any mention of prerequisites, context (e.g., for business reporting), or comparisons to similar tools like 'balance-sheet' or 'cash-flow-statement' in the sibling list. The only additional context is '(Browser-based tool)', which is vague and does not inform usage decisions.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
pro-mattingPro MattingBInspect
Professional-grade AI matting with fine hair-level edge detection. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions 'Professional-grade AI matting' and 'Browser-based,' but lacks details on input requirements (e.g., image formats), output behavior, processing time, or any limitations. This leaves significant gaps for an AI agent to understand how to invoke it effectively.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise, using only two phrases that directly convey the core functionality and platform. Every word earns its place, with no redundant information, making it easy to parse quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity of an AI matting tool with no annotations, no output schema, and 0 parameters, the description is incomplete. It lacks details on what the tool expects as input (e.g., image upload), what it returns, error handling, or performance traits, leaving the AI agent with insufficient context for reliable use.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description doesn't add param info, which is appropriate, but it could hint at implicit inputs (e.g., image data). Since the baseline for 0 params is 4, this score reflects that the description doesn't detract from an already minimal schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose as 'Professional-grade AI matting with fine hair-level edge detection,' specifying the action (matting) and key capability (hair-level edge detection). It distinguishes itself from sibling tools like 'bg-remover' by emphasizing precision and professional quality. However, it could be more specific about the exact output or use cases.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives, such as 'bg-remover' or other image-processing tools in the sibling list. It mentions it's 'Browser-based,' which hints at accessibility but doesn't clarify scenarios or prerequisites for selection.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
purchase-cost-manager仕入れ原価管理表BInspect
仕入れ先×商品の原価を記録。粗利率を自動計算 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full disclosure burden. It successfully indicates two key behavioral traits: automatic calculation of gross margins and that it is browser-based (UI-driven). However, it omits critical behavioral context such as data persistence (where records are stored), idempotency, side effects of repeated calls, or return value format.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is exceptionally concise with two clear sentences plus a parenthetical modifier. It front-loads the core value proposition (cost recording) before the secondary feature (margin calculation) and ends with the UI classification. No words are wasted.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given zero input parameters and no output schema, the description provides the minimum viable context for a browser-based utility, explaining what functional capability the user gets (cost management with auto-calculation). However, it lacks closure on what the tool returns (e.g., a URL, a file, or a confirmation message) which is critical information when the schema provides no structural hints.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema contains zero parameters. Per the rubric, zero-parameter tools receive a baseline score of 4. The description implicitly hints at the data model (supplier × product matrix) but does not need to compensate for missing schema documentation since there are no parameters to describe.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool records purchase costs by supplier and product (仕入れ先×商品の原価を記録) and automatically calculates gross margins (粗利率を自動計算). It identifies the specific resource managed (purchase costs) and the core action (recording/calculation), though it does not explicitly differentiate from similar calculator siblings like shopify-profit-calculator or amazon-fba-calculator.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides minimal usage guidance. While the '(Browser-based tool)' parenthetical hints at the interaction model (opening a web UI versus returning raw data), there is no explicit guidance on when to use this versus alternatives, prerequisites for supplier/product data, or expected workflow.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
purchase-orderPurchase OrderCInspect
Purchase orders with delivery terms and tax calculation. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions 'delivery terms and tax calculation' but doesn't clarify if this tool creates, modifies, or analyzes purchase orders, what inputs it requires (despite having 0 parameters), or what outputs to expect. The phrase 'Browser-based tool' is irrelevant to behavioral traits.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is brief but inefficiently structured. The first part 'Purchase orders with delivery terms and tax calculation' is somewhat informative, but the parenthetical '(Browser-based tool)' adds no value for an AI agent. It's not front-loaded with critical usage information.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no annotations, no output schema, and 0 parameters, the description is incomplete. It doesn't explain what the tool does operationally (e.g., generates a PDF, calculates costs), what results to expect, or how it differs from similar tools. For a tool with no structured data, more descriptive context is needed.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
With 0 parameters and 100% schema description coverage, the baseline is high. The description adds no parameter information, which is acceptable since there are no parameters to document. It doesn't compensate for any gaps because none exist.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description 'Purchase orders with delivery terms and tax calculation' states what the tool handles but is vague about its specific action. It doesn't specify a verb (e.g., create, generate, calculate) or distinguish it from siblings like 'invoice-generator' or 'purchase-cost-manager'. The phrase 'Browser-based tool' is an implementation detail, not a purpose statement.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives. With siblings like 'invoice-generator' and 'purchase-cost-manager', the description offers no context for selection, prerequisites, or exclusions. The user must guess based on the tool name alone.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
qr-designerQR Code DesignerBInspect
Create styled QR codes with custom logos and colors. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It mentions 'Browser-based tool', which adds some context about the execution environment. However, it lacks details on behavioral traits such as whether it requires internet access, how it handles errors, what the output format is (e.g., image file, URL), or any rate limits. For a tool with zero annotation coverage, this is insufficient.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise: two short phrases that directly state the tool's function and environment. Every word earns its place with no redundancy or fluff. It's front-loaded with the core purpose.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool has 0 parameters, no annotations, and no output schema, the description is minimally adequate. It covers the basic purpose and environment but lacks details on output format, error handling, or usage context. For a tool with no structured data, it should do more to be fully complete, but it's not completely inadequate.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description doesn't add parameter details, which is appropriate. A baseline of 4 is given because the schema fully covers the parameters (none), and the description doesn't need to compensate.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Create styled QR codes with custom logos and colors.' It specifies the action (create), resource (QR codes), and key features (styled, custom logos/colors). However, it doesn't differentiate from sibling tools like 'barcode-generator' or 'stamp-maker' that might also generate visual codes or images, so it's not a perfect 5.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It doesn't mention prerequisites, ideal scenarios, or comparisons to siblings like 'barcode-generator' or 'image-resizer'. The only contextual note is 'Browser-based tool', which hints at the environment but not usage decisions.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
quote-generatorQuote GeneratorBInspect
Create quotes with auto tax calculation and PDF export. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It mentions 'auto tax calculation' and 'PDF export', which are useful behavioral traits, but lacks details on permissions, rate limits, side effects (e.g., whether quotes are saved), or error handling. For a tool with zero annotation coverage, this is insufficient.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is highly concise and front-loaded: a single sentence that efficiently conveys the core functionality and context ('Browser-based tool'). Every word earns its place, with no redundancy or fluff.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool has 0 parameters, no annotations, and no output schema, the description is adequate but minimal. It covers the basic purpose and key features but lacks depth on behavioral aspects (e.g., how PDF export works) and output details. For a tool with no structured data, it meets the minimum viable threshold.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% description coverage, so no parameter documentation is needed. The description appropriately avoids discussing parameters, focusing instead on functionality. A baseline of 4 is applied since the schema fully covers the absence of parameters.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Create quotes with auto tax calculation and PDF export.' It specifies the action (create), resource (quotes), and key features (auto tax calculation, PDF export). However, it doesn't explicitly differentiate from sibling tools like 'invoice-generator' or 'receipt-generator', which might have overlapping functionality, preventing a perfect score.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides minimal usage guidance. It mentions the tool is 'Browser-based', which hints at its context, but offers no explicit advice on when to use it versus alternatives (e.g., 'invoice-generator' for invoices vs. quotes), prerequisites, or exclusions. This leaves the agent with little direction.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
rakuten-banner-maker楽天バナーメーカーAInspect
お買い物マラソン/スーパーSALE用バナーをテンプレから即生成 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Adds crucial 'Browser-based tool' disclosure indicating interactive UI behavior, supplementing the absent annotations. However, lacks details on return values, side effects, or what happens after generation (file download vs URL return).
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Single compact Japanese sentence plus parenthetical note. Every element serves a purpose: target event, resource, method (templates), and interaction mode (browser). Zero redundancy.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Adequate for tool identification given zero parameters, but lacks behavioral specifics (e.g., whether it returns a download link, opens external URL, or requires user interaction) that annotations or output schema would normally provide.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Baseline 4 for zero-parameter tools. No parameters require semantic explanation beyond the empty schema, and description appropriately focuses on tool behavior rather than inventing parameter documentation.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description uses specific verb '生成' (generate) + resource 'バナー' (banner) and distinguishes from generic banner tools by specifying exclusive use for 'お買い物マラソン/スーパーSALE' (Rakuten-specific shopping events). Clearly differentiates from sibling calculators and generic image tools.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Provides clear usage context by specifying exact scenarios (Shopping Marathon/Super SALE). While it doesn't explicitly name alternatives, the specific event targeting gives agents clear selection criteria without exclusions.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
rakuten-coupon-simulator楽天クーポン設計ツールBInspect
割引率/条件→利益シミュレーション。最適なクーポン設計を支援 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Annotations are absent, so description carries full burden. It adds '(Browser-based tool)' which signals UI interaction rather than API calculation, but lacks disclosure of what gets returned to the agent (URL? simulation results?), expected execution time, or whether results are storable.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely efficient: two short sentences front-loaded with core value proposition (simulation) and execution mode (browser-based). No redundant text.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given zero parameters and no output schema, the description covers the functional concept but omits critical execution details—specifically what the agent receives after invoking a browser-based tool (e.g., simulation results, URL, or instructions). 'Optimal design support' is vague regarding tangible outputs.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Zero parameters present. Per scoring rubric, 0 params = baseline 4. Description provides no parameter details, but none are needed given empty schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Clear specific verb (シミュレーション/simulation) and resource (クーポン/coupon + 利益/profit). The description specifies inputs (割引率/条件/discount rates/conditions) and outputs (profit simulation), distinguishing it from sibling Rakuten tools like fee-calculator and rpp-calculator which handle different business logic.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance on when to use this versus other Rakuten commerce tools (e.g., fee-calculator, pricing-guide). Only behavioral note is '(Browser-based tool)' which hints at execution mode but doesn't specify prerequisites, eligibility, or decision criteria.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
rakuten-fee-calculator楽天市場手数料計算機AInspect
システム利用料+決済手数料+ポイント→利益を正確計算 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Without annotations, the description carries full burden. It successfully discloses the browser-based nature and the specific fee components included in calculations. However, it lacks detail on how the browser interaction works (does it return a URL? require user input?) and omits any mention of data sources or authentication requirements.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely concise: the formula notation (A+B+C→result) efficiently conveys the calculation scope in minimal characters, and the 'Browser-based tool' parenthetical adds crucial behavioral context without verbosity. Every element earns its place.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the absence of parameters, annotations, and output schema, the description provides minimal viable context for a simple calculator tool. However, it fails to clarify the browser interaction model or what return value the agent should expect, leaving gaps in contextual completeness.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
With zero parameters, the baseline is 4. The description adds value by conceptually outlining what inputs the calculator considers (system fees, payment fees, points) even though these aren't formal schema parameters.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the specific action (accurate profit calculation) and the specific resource (Rakuten Market fees: system usage fee + payment fee + points). It effectively distinguishes from sibling calculators like amazon-fba-calculator or mercari-calculator by naming Rakuten-specific fee components.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No explicit guidance provided on when to use this versus other marketplace calculators (like yahoo-shopping-calculator or shopify-profit-calculator). The 'Browser-based tool' note hints at behavioral differences but doesn't constitute usage guidelines.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
rakuten-rpp-calculator楽天RPP広告計算機AInspect
RPP広告費→CPC/ROAS/損益分岐を計算 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Discloses 'Browser-based tool' which indicates interactive UI vs direct API return, adding crucial behavioral context. However, with zero annotations, description carries full burden and lacks safety profile (read-only vs destructive), authentication requirements, or rate limits.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Single efficient sentence with parenthetical note. Front-loaded with calculation purpose (RPP広告費→metrics). Every element earns its place including the Japanese text which matches the localized title.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a zero-parameter tool without output schema, description adequately covers the calculation scope. However, gaps remain: does not clarify if it returns a URL, opens a browser, or returns raw values; omits prerequisites for the RPP ad cost data.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema contains zero parameters. Per rubric, 0 params = baseline 4. Description appropriately does not invent parameters, though it could explicitly state 'no parameters required' to clarify the empty schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Specific verb (計算/calculate) + resource (RPP広告費) + clear outputs (CPC/ROAS/損益分岐). Title and Japanese description precisely distinguish this from fee calculators (rakuten-fee-calculator) and banner tools (rakuten-banner-maker) by specifying RPP advertising metrics.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance on when to use versus sibling calculators (amazon-acos-calculator, yahoo-shopping-calculator) or prerequisites for RPP ad data. No mention of required inputs or setup steps despite being a browser-based interactive tool.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
receipt-generatorReceipt GeneratorBInspect
Invoice-compliant receipts with auto tax calculation. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It mentions 'browser-based tool', hinting at a UI component, but doesn't disclose critical behavioral traits: whether it's interactive, if it requires user input beyond the tool call, what happens after generation (e.g., download, display), or any limitations like file formats or tax jurisdictions. For a tool with zero annotation coverage, this is inadequate.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise and front-loaded: the first part states the core functionality, and the second adds a useful technical note. Every sentence earns its place, with no wasted words or redundancy. It's appropriately sized for a simple tool.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (likely moderate, involving tax calculation and compliance), lack of annotations, and no output schema, the description is incomplete. It doesn't explain what the tool returns (e.g., a PDF, HTML, or data), how tax calculations work, or any error conditions. For a tool with behavioral and output uncertainties, this is insufficient.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The tool has 0 parameters, with 100% schema description coverage (since there are no parameters to describe). The description doesn't need to add parameter semantics, so a baseline of 4 is appropriate—it's not a 5 because it doesn't explicitly state 'no parameters required', but the absence of parameters is handled well by the schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: generating invoice-compliant receipts with auto tax calculation. It specifies the verb ('generates') and resource ('receipts'), and distinguishes itself from siblings like 'invoice-generator' by focusing on receipts rather than invoices. However, it doesn't explicitly differentiate from other receipt-related tools (none in the list), so it's not a perfect 5.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It mentions 'browser-based tool' as a technical note, but offers no context about prerequisites, typical use cases, or comparisons with siblings like 'invoice-generator' or 'expense-report'. This leaves the agent without clear usage direction.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
recruitment-fee-calcRecruitment Fee CalcCInspect
Staffing agency fee, refund terms, and KPI back-calculation.
| Name | Required | Description | Default |
|---|---|---|---|
| feeRate | No | Fee rate (%) | |
| feeType | No | Fee type | percentage |
| fixedFee | No | Fixed fee in JPY (when feeType=fixed) | |
| monthlyTarget | No | Monthly target placements | |
| averagePlacementFee | No | Average placement fee in JPY | |
| theoreticalAnnualSalary | Yes | Theoretical annual salary in JPY |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It mentions 'refund terms' and 'KPI back-calculation' but does not explain what these entail operationally (e.g., calculation methods, output format, or any side effects like data persistence). This leaves key behavioral traits undefined for a tool with multiple parameters.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient phrase with three key components, making it front-loaded and concise. However, it could be more structured by clarifying the tool's action (e.g., 'Calculate staffing agency fees, refund terms, and KPI back-calculation').
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (6 parameters, no output schema, no annotations), the description is incomplete. It does not explain what the tool returns (e.g., fee amount, refund calculations, KPI metrics) or how the inputs interact, leaving significant gaps for an AI agent to understand the tool's full context and usage.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so parameters are well-documented in the schema itself (e.g., 'feeRate' as 'Fee rate (%)'). The description adds no additional meaning beyond the schema, such as explaining relationships between parameters (e.g., how 'feeType' affects calculations). Baseline score of 3 is appropriate as the schema handles parameter semantics adequately.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description 'Staffing agency fee, refund terms, and KPI back-calculation' lists three components but lacks a clear verb-action pairing (e.g., 'calculate' or 'compute') and does not specify how these components relate or what the tool outputs. It distinguishes from siblings by focusing on recruitment fees, but the purpose remains vague without explicit functionality.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives (e.g., other fee calculators like 'base-fee-calculator' or 'rakuten-fee-calculator'), nor any context about prerequisites or typical scenarios. The description implies usage for staffing agency contexts but offers no explicit when/when-not instructions.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
regex-testerRegex TesterCInspect
Real-time regex testing and debugging.
| Name | Required | Description | Default |
|---|---|---|---|
| flags | No | Regex flags (g, i, m, s, u, y) | g |
| pattern | Yes | Regular expression pattern | |
| testString | Yes | String to test against the pattern |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It mentions 'real-time' testing, implying immediate feedback, but doesn't disclose behavioral traits like error handling, output format, rate limits, or side effects. For a tool with no annotations, this is insufficient.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient phrase with no wasted words. It's appropriately sized and front-loaded, making it easy to parse quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no annotations and no output schema, the description is incomplete. It doesn't explain what the tool returns (e.g., matches, groups, errors) or behavioral details, which is inadequate for a tool that performs regex operations.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema fully documents parameters (pattern, testString, flags). The description adds no additional parameter semantics beyond what's in the schema, resulting in the baseline score of 3.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description 'Real-time regex testing and debugging' clearly states the tool's function (testing and debugging regular expressions) but is somewhat vague about the specific actions. It doesn't distinguish from siblings, though no obvious regex-related siblings exist in the list.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance on when to use this tool versus alternatives is provided. The description doesn't mention prerequisites, context, or exclusions, leaving the agent to infer usage based on the tool name alone.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
resident-tax-calc住民税シミュレーターAInspect
住民税の年額・月額を瞬時に計算。所得割・均等割の内訳表示 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Discloses that it is a 'Browser-based tool', which is crucial context given the empty input schema, indicating an interactive UI rather than a parameterized API. Also mentions the breakdown display format. No annotations exist to contradict.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely concise two-part structure: first clause states the calculation action, second states the breakdown display. The 'Browser-based' parenthetical efficiently explains the interaction model. Zero wasted words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a zero-parameter calculation tool with no output schema, the description adequately explains the function, output format (annual/monthly, detailed breakdown), and interaction mode (browser-based). Sufficient for tool selection.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema has 0 parameters, establishing a baseline of 4 per scoring rules. The description appropriately does not mention parameters since none exist.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Specifically states it calculates resident tax (住民税) annual and monthly amounts, and displays breakdown of income levy (所得割) and equal levy (均等割). The specific tax terminology clearly distinguishes it from sibling tools like freelance-tax-calc, withholding-tax-calc, and tax-return-calc.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Provides clear context through specific naming (resident tax vs other taxes) and technical components (所得割・均等割), allowing the agent to distinguish it from other tax calculators. However, lacks explicit 'when not to use' or alternative references.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
resignation-generatorResignation LetterBInspect
Proper-format resignation letters in 30 seconds. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It mentions 'browser-based tool,' implying it runs in a browser environment, but doesn't disclose critical behavioral traits like whether it requires user input beyond the interface, how it handles data privacy, or what the output looks like (e.g., file format). This leaves significant gaps for a tool that generates documents.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise and front-loaded: it states the main function ('Proper-format resignation letters'), a key feature ('in 30 seconds'), and a technical note ('Browser-based tool') in just two phrases. Every part earns its place with no wasted words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool has 0 parameters and no annotations or output schema, the description is minimally complete. It covers what the tool does and its environment, but lacks details on output format, customization options, or usage context, which are important for a document generator. It's adequate but has clear gaps.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description doesn't add parameter details, which is appropriate here. A baseline of 4 is given as it compensates adequately for the lack of parameters by focusing on the tool's core function.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: generating 'proper-format resignation letters' with a time estimate ('in 30 seconds'). It specifies the output format and speed, though it doesn't explicitly differentiate from sibling tools like 'resume-generator' or 'career-history-generator' beyond the specific document type.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It mentions it's 'browser-based,' which is a technical constraint, but offers no context about prerequisites, ideal scenarios, or comparisons to similar tools in the list (e.g., 'contract-generator' or 'terms-generator').
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
resume-generatorResume GeneratorBInspect
JIS-standard Japanese resume (rirekisho) with PDF export. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It mentions PDF export and browser-based nature, which adds some behavioral context (output format and interface). However, it lacks details on permissions, rate limits, error handling, or what the generated PDF contains (e.g., template details). For a tool with zero annotation coverage, this is insufficient.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is concise and front-loaded, stating the core function in the first part. The 'browser-based tool' note adds useful context without verbosity. However, it could be slightly more structured by separating key features, but it's efficient overall.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool has 0 parameters, no annotations, and no output schema, the description is minimally complete. It covers the purpose and output format but lacks details on behavioral traits, usage scenarios, or error handling. For a simple tool, this might suffice, but gaps remain in transparency and guidelines.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The tool has 0 parameters, and schema description coverage is 100%, so no parameter documentation is needed. The description doesn't add parameter semantics, but with no parameters, the baseline is high. It implicitly suggests the tool might generate resumes without user input, which is adequate.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's function: generating a JIS-standard Japanese resume (rirekisho) with PDF export. It specifies the resource (resume) and format (PDF), and distinguishes it from siblings like 'career-history-generator' by mentioning the Japanese standard. However, it doesn't explicitly contrast with 'resignation-generator' or other document tools, keeping it from a perfect score.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides minimal usage guidance, only noting it's 'browser-based,' which implies a web interface but doesn't specify when to use this tool over alternatives like 'career-history-generator' or 'resignation-generator.' No explicit when-to-use or when-not-to-use instructions are given, leaving the agent to infer context.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
review-reply-templateレビュー返信テンプレートAInspect
高評価/低評価への返信パターンを自動生成。炎上防止 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full disclosure burden. It adds the '炎上防止' behavioral trait (de-escalation focus) and notes '(Browser-based tool)' suggesting execution context. However, it lacks details on output format (templates vs. opened browser), persistence, or whether human-in-the-loop confirmation is expected before sending replies.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely efficient two-clause Japanese sentence. Information density is high: target domain (high/low ratings), action (auto-generate), deliverable (reply patterns), and quality attribute (flamewar prevention) are all front-loaded with zero redundancy.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given zero input parameters and the tool's apparent simplicity, the description adequately covers the value proposition and scope. The 'Browser-based' parenthetical partially addresses the missing output schema by hinting at the interaction model, though explicit return value documentation would strengthen this.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema contains zero parameters. Per calibration rules, 0 params establishes a baseline of 4. The description appropriately requires no additional parameter semantics since there are no inputs to document.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description clearly states the specific action (auto-generate reply patterns), target resource (high/low ratings), and unique value proposition (flamewar prevention). It effectively distinguishes this from sibling content generators like article-outline-generator by specifying the review-reply domain and risk-mitigation purpose.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The phrase '炎上防止' (flamewar prevention) implies the tool should be used when reply tone management is critical, providing implicit context. However, there is no explicit guidance on when to use this versus general text generation tools, or prerequisites like requiring actual review content.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
rewrite-assistantリライト支援ツールBInspect
古い記事の改善ポイントを自動検出。情報鮮度・構成・重複をチェック (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Discloses 'Browser-based tool' execution mode and scope of analysis (freshness, structure, duplication). However, with no annotations provided, the description omits critical behavioral details: output format (returns report vs opens UI?), whether it modifies input, side effects, or completion criteria.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely efficient: two sentences covering purpose and scope, plus parenthetical execution context. Every element earns its place with no redundancy or filler content. Front-loaded with core function stated immediately.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Adequate for a zero-parameter browser tool, covering intent and execution environment. However, lacks output specification (crucial given no output schema) and bite-size completeness for the claimed analysis function.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Zero parameters present per input schema. Description correctly avoids parameter documentation since none exist. Baseline score applies as no compensation is required.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Clear specific verb (検出/detects) and resource (古い記事/old articles) with concrete aspects checked (freshness, structure, duplication). Distinguishes from outline generators by focusing on analysis of existing content rather than creation. However, it fails to differentiate from similar analysis sibling 'sales-writing-analyzer'.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Only provides implicit context (for 'old articles'). No explicit when-to-use guidance, prerequisites, or alternatives mentioned despite numerous sibling content tools (article-outline-generator, sales-writing-analyzer) where selection criteria would be valuable.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
salary-vs-freelance会社員vsフリーランス比較BInspect
税金・保険・手取りを完全比較。損益分岐点を自動算出 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full disclosure burden. It notes the tool is '(Browser-based tool)', which hints at interactive/UI behavior rather than API data return, but does not explain what this means for the user experience (e.g., opens a browser window, returns a URL, requires human interaction). It mentions the calculation scope but omits data sources, persistence, or side effects.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise with two information-dense sentences. Key actions (比較, 算出) are front-loaded, and the parenthetical '(Browser-based tool)' efficiently signals implementation type without verbosity. No filler text is present.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the absence of annotations and output schema, the description adequately covers the tool's purpose but leaves ambiguity around the 'browser-based' execution model and what exactly is returned to the agent. For a financial calculation tool, it lacks context on input requirements (implied interactive) or result format completeness.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has zero properties. According to scoring rules, 0 parameters warrant a baseline score of 4. The description does not contradict this empty schema, though it does not explicitly confirm that the tool operates without inputs (e.g., via interactive browser UI) either.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool compares taxes, insurance, and net pay between employment types (会社員vsフリーランス) and calculates break-even points (損益分岐点). It uses specific verbs (比較, 算出) and identifies the resource domain. However, it does not explicitly differentiate from siblings like 'freelance-tax-calc' or 'take-home-pay-calc' which cover overlapping domains.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It does not indicate that this is specifically for side-by-side comparison scenarios, or when to prefer 'freelance-tax-calc' for single-status calculations instead. No prerequisites or exclusion criteria are mentioned.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
sales-writing-analyzerセールスライティング分析BInspect
PASONA/AIDMA構成を自動判定。成約率を上げる改善点を提示 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full disclosure burden. It mentions '(Browser-based tool)' indicating the runtime environment, but fails to state whether the operation is read-only, what data persists, or any rate limiting. It implies analysis is non-destructive but doesn't confirm safety characteristics.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The Japanese description is exceptionally compact—two short sentences delivering the core value proposition without redundancy. Every element (framework names, outcome, browser indicator) earns its place with no filler content.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the lack of output schema and zero input parameters, the description should explain the interaction model (how text is submitted and results returned). While the PASONA/AIDMA framework reference provides domain context, the operational specifics remain ambiguous for an AI agent attempting invocation.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema contains zero parameters. Per the evaluation rules, zero parameters establishes a baseline score of 4. The description does not need to compensate for missing schema documentation, though it could have clarified the input mechanism (e.g., clipboard-based) given the empty parameter set.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly identifies the tool's function using specific marketing frameworks (PASONA/AIDMA) and states it suggests improvements for conversion rates. However, it lacks clarity on how input is provided given the zero-parameter schema—whether it analyzes clipboard content, opens an interactive browser window, or requires pre-selected text.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus siblings like 'rewrite-assistant' (for editing) or 'article-outline-generator' (for planning). There are no mentions of prerequisites, such as needing sales copy text prepared, or exclusions for when this analysis is inappropriate.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
seating-chartSeating ChartCInspect
Random seating with constraints. Printable PDF. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It mentions 'random seating with constraints' and 'printable PDF', but lacks details on behavioral traits such as what constraints are supported, whether it requires user input or generates automatically, if it has rate limits, or how the PDF is delivered. This leaves significant gaps for a tool that likely involves generation and output handling.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise with three brief phrases separated by periods, front-loading key actions ('Random seating with constraints. Printable PDF.') and ending with context ('Browser-based tool'). Every part adds value without waste, making it efficient and well-structured.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity of generating seating charts with constraints and PDF output, the description is incomplete. No annotations or output schema exist to clarify behavior or results. The description lacks details on constraints, PDF generation specifics, or usage context, making it inadequate for an agent to understand the tool's full scope and limitations.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, meaning no parameters are documented in the schema. The description does not mention any parameters, which is appropriate here. It implies the tool might operate without inputs or with default settings, but does not clarify if parameters are hidden or unnecessary. Baseline is 4 for 0 parameters, as no additional param info is needed.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description states the tool generates a 'random seating' arrangement with 'constraints' and outputs a 'printable PDF', which clarifies the action and resource. However, it does not differentiate from sibling tools beyond being 'browser-based', which is vague and not unique among the many document/PDF generators in the list.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives. The description mentions it is 'browser-based', but this does not help in choosing between similar tools like 'certificate-maker' or 'receipt-generator' that might also produce PDFs. There are no explicit when/when-not instructions or named alternatives.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
shipping-calculatorShipping CalculatorBInspect
Compare rates across Yamato, Yu-Pack, and Sagawa. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It mentions the tool is 'Browser-based', which suggests it might open a web interface or require browser interaction, but doesn't detail what the tool actually does (e.g., opens a calculator UI, fetches live rates, simulates costs) or any constraints like rate limits, permissions, or data sources.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise—two short phrases with zero wasted words. It front-loads the core purpose and adds a brief technical note, making it easy to scan and understand quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the lack of annotations and output schema, the description is incomplete. It doesn't explain what the tool returns (e.g., a comparison table, best option, detailed breakdowns) or how it behaves (e.g., interactive vs. static). For a tool with potential complexity in shipping rate comparisons, more context on functionality and results is needed.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description doesn't add parameter details, which is appropriate here, earning a baseline score of 4 for not introducing unnecessary information.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Compare rates across Yamato, Yu-Pack, and Sagawa.' It specifies the verb ('compare') and the resource ('rates'), and distinguishes the three specific carriers. However, it doesn't explicitly differentiate from sibling tools like 'mercari-shipping-compare', which might serve a similar function in a different context.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives, such as other shipping calculators or comparison tools in the sibling list. The parenthetical note '(Browser-based tool)' hints at a technical context but doesn't clarify functional use cases or prerequisites.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
shopify-profit-calculatorShopify利益計算機BInspect
プラン別月額+決済手数料→損益分岐点と利益を計算 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Discloses the calculation logic (plan fees + payment fees → breakeven/profit) and notes it is 'Browser-based tool', hinting at execution environment. However, with no annotations, it lacks details on side effects, data persistence, or whether this opens an interactive UI versus returning static data.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely concise Japanese description using efficient arrow notation (→) to show input-output relationship. The parenthetical '(Browser-based tool)' adds execution context without verbosity. Zero wasted words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Adequate for a zero-parameter tool but leaves ambiguity: the description references specific inputs (plan-specific fees) that don't appear in the empty parameter schema, creating confusion about whether user input is required via browser or handled automatically. Doesn't clarify output format.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema has zero parameters, which per guidelines receives a baseline of 4. The description conceptually mentions inputs (plan fees, transaction fees) but these are not exposed as API parameters, suggesting the tool may be interactive or use internal defaults.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Clearly states it calculates break-even point and profit using Shopify plan-specific monthly fees and payment processing fees. The specific calculation methodology (breakeven + profit) and resource (Shopify plans) are well-defined, though it doesn't explicitly differentiate from the generic 'profit-loss' sibling tool.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Provides no guidance on when to use this versus sibling e-commerce calculators like amazon-fba-calculator, rakuten-fee-calculator, or yahoo-shopping-calculator. No mention of prerequisites or when not to use.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
similarity-checker類似度チェッカーAInspect
テキスト間の類似度をn-gramで判定。サーバー送信ゼロ (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description carries full burden. Discloses execution model (browser-based, zero server transmission) and algorithm (n-gram). Missing output format details but covers critical privacy behavior.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two concise fragments: first states purpose, second states execution model. No redundancy, front-loaded with essential information.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Functionally complete for a zero-parameter tool, but lacks description of return values/similarity score format. Absence of output schema makes this gap noticeable.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Zero parameters present; per guidelines, baseline score is 4. No parameter documentation required.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Specific verb (判定/determine), resource (テキスト間の類似度/text similarity), and method (n-gram) clearly stated. Distinguishes from sibling diff-checker by specifying n-gram algorithm.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Implicit guidance provided via 'サーバー送信ゼロ' (zero server transmission), indicating suitability for private/sensitive data. Lacks explicit when-not-to-use or alternative comparisons.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
sku-generatorSKUジェネレーターCInspect
商品名・カテゴリ・サイズ・色からSKUコードを自動生成 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure but only adds the 'Browser-based' context without explaining side effects, data persistence, or how the four mentioned inputs are handled given the empty parameter schema. It fails to clarify the execution model or why inputs described are not present in the schema.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence that front-loads the core functionality with zero redundant words. The parenthetical '(Browser-based tool)' appropriately qualifies the execution context without cluttering the main functional statement.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the critical discrepancy between the four described inputs and the empty input schema, the description fails to provide sufficient context for correct invocation. The lack of output schema and annotations compounds the incompleteness, leaving the agent uncertain how to supply the required product data or what to expect in return.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
While the baseline for zero parameters is 4, the description lists four specific input fields (product name, category, size, color) that do not exist in the empty input schema, creating a confusing mismatch for the agent. It adds semantic meaning about expected data but contradicts the structured schema definition that shows zero properties.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool generates SKU codes ('SKUコードを自動生成') from specific inputs including product name, category, size, and color, distinguishing it from sibling generators like barcode-generator or uuid-generator. The verb and resource are specific and actionable.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides minimal guidance, only noting it is 'Browser-based' without explaining when to use this tool versus alternatives like barcode-generator, or prerequisites for invocation. There is no explanation of the browser-based workflow or when programmatic SKU generation is appropriate.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
smart-resizeSmart ResizeBInspect
Batch image resizing with AI subject detection and smart crop. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It mentions 'batch image resizing' and 'AI subject detection,' but lacks critical behavioral details: what inputs it accepts (e.g., file types, size limits), whether it modifies originals, authentication needs, rate limits, or output format. For a tool with AI features and no annotations, this is a significant gap.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise: one sentence with no wasted words. It front-loads the core functionality ('Batch image resizing with AI subject detection and smart crop') and adds a brief context note ('Browser-based tool'). Every part earns its place.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity (AI-based image processing), lack of annotations, no output schema, and 0 parameters, the description is incomplete. It doesn't explain how the tool works, what it returns, or any behavioral constraints. For a tool with advanced features like AI detection, more context is needed to guide an AI agent effectively.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameters need documentation. The description adds value by hinting at implicit inputs ('batch image'), but since there are no parameters, the baseline is 4. It doesn't fully explain how inputs are provided (e.g., via UI), so it's not a 5.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Batch image resizing with AI subject detection and smart crop.' It specifies the verb ('resizing'), resource ('image'), and key features ('batch,' 'AI subject detection,' 'smart crop'). However, it doesn't explicitly differentiate from sibling tools like 'image-resizer' or 'image-crop,' which is why it's not a 5.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It mentions it's 'Browser-based,' but this doesn't help an AI agent choose between this and other image tools like 'image-resizer' or 'image-crop.' There are no explicit when/when-not instructions or named alternatives.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
sns-post-templateSNS投稿文テンプレAInspect
動画告知/切り抜き紹介のSNS投稿文をプラットフォーム別に生成 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description carries full burden. Includes parenthetical note '(Browser-based tool)' which hints at execution context, but lacks details on output format, supported platforms, or side effects. Adequate but minimal behavioral disclosure.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Single sentence with front-loaded purpose. Every element serves a function: main clause declares capability, parenthetical discloses execution model. Zero waste in Japanese text.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a zero-parameter tool with no output schema, the description adequately covers the functional purpose. However, lacks return value documentation or specific platform enumeration, leaving some uncertainty about expected output.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema contains 0 parameters. With zero parameters, baseline score is 4 according to rubric (no parameters to document).
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Clear specific verb (生成/generate), resource (SNS投稿文/SNS post text), and scope (動画告知/切り抜き紹介 for video announcements/clips, プラットフォーム別/by platform). Distinguishes from siblings like youtube-description-generator or title-multilingual by focusing on cross-platform SNS post content.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Provides implied usage through specific content scope (video announcements and clip introductions rather than general posts). However, lacks explicit when-to-use guidance or named alternatives from the extensive sibling list.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
sql-formatterSQL FormatterCInspect
Format and beautify SQL queries.
| Name | Required | Description | Default |
|---|---|---|---|
| sql | Yes | SQL query to format | |
| minify | No | Minify instead of format | |
| dialect | No | SQL dialect | standard |
| indentStyle | No | Indent style | 2spaces |
| keywordCase | No | Keyword case | upper |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It states the tool formats and beautifies SQL queries, implying a read-only transformation, but doesn't cover aspects like error handling, performance, rate limits, or output format. For a tool with no annotations, this leaves significant gaps in understanding its behavior.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise and front-loaded with a single, clear sentence: 'Format and beautify SQL queries.' There is no wasted verbiage or unnecessary elaboration, making it efficient and easy to parse.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (5 parameters, no annotations, no output schema), the description is insufficient. It doesn't explain what the formatted output looks like, how errors are handled, or any behavioral nuances. While the schema covers parameters, the lack of output schema and annotations means the description should do more to complete the context, which it fails to do.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The description adds no parameter-specific information beyond the input schema, which has 100% coverage with detailed descriptions for all 5 parameters. Since the schema fully documents parameters like 'sql', 'minify', and 'dialect', the description doesn't need to compensate, but it also doesn't provide additional context or examples, resulting in a baseline score.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Format and beautify SQL queries.' It specifies the verb ('format and beautify') and resource ('SQL queries'), making the function unambiguous. However, it doesn't explicitly differentiate from sibling tools like 'code-formatter' or 'json-expert', which handle other types of formatting, so it doesn't reach the highest score.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It doesn't mention sibling tools like 'code-formatter' for non-SQL code or specify contexts where SQL formatting is preferred over other operations. There's no indication of prerequisites or exclusions, leaving usage entirely implicit.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
stamp-makerStamp MakerBInspect
Create digital hanko seals (personal, corporate, date stamps). (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It states the tool creates digital hanko seals and is browser-based, but lacks details on permissions, rate limits, output format, or any side effects. For a creation tool with zero annotation coverage, this is insufficient to inform safe and effective use.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise and front-loaded: it states the core purpose in the first part and adds a brief contextual note ('Browser-based tool') in parentheses. Every word contributes essential information without waste, making it efficient and easy to parse.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool has 0 parameters and no output schema, the description is minimally complete for its simplicity. It covers what the tool does and the execution environment. However, as a creation tool with no annotations, it lacks details on behavioral aspects like output format or error handling, which could be important for user expectations.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, meaning no parameters are documented in the schema. The description doesn't mention any parameters, which is appropriate here as there are none to explain. It avoids redundancy and correctly aligns with the schema, earning a high score for semantic alignment in a parameterless context.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Create digital hanko seals (personal, corporate, date stamps).' It specifies the action ('Create') and the resource ('digital hanko seals'), with examples of types. However, it doesn't explicitly differentiate from sibling tools like 'pdf-stamper' or 'certificate-maker', which might involve similar stamping or creation concepts, leaving room for ambiguity in sibling context.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It mentions it's a 'Browser-based tool', which hints at the execution environment but doesn't specify use cases, prerequisites, or exclusions. Without explicit when-to-use or when-not-to-use instructions, users must infer usage from the purpose alone.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
string-transformString TransformBInspect
Transform strings between camelCase, snake_case, kebab-case, slug, etc.
| Name | Required | Description | Default |
|---|---|---|---|
| mode | No | Transform mode | case-convert |
| input | Yes | String to transform |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It states the tool transforms strings but does not clarify output format, error handling, or limitations (e.g., character constraints, performance). For a tool with no annotations, this leaves significant gaps in understanding how the tool behaves beyond basic functionality.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence that directly states the tool's function without unnecessary words. It is front-loaded with the core action ('Transform strings') and includes relevant examples, making it easy to grasp quickly. Every part of the sentence contributes to clarity.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's low complexity (2 parameters, no output schema, no annotations), the description is minimally adequate. It covers the basic purpose but lacks details on usage, behavioral traits, and output expectations. Without annotations or an output schema, the description should provide more context to be fully complete, but it meets the bare minimum for a simple utility tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents both parameters ('mode' with enum values and 'input' as a string). The description lists transformation types (camelCase, snake_case, etc.) that align with the 'mode' enum, adding minimal context. Since the schema provides full coverage, the baseline score of 3 is appropriate, as the description does not significantly enhance parameter understanding.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: transforming strings between various case formats (camelCase, snake_case, kebab-case, slug, etc.). It specifies the verb 'transform' and the resource 'strings', making the intent unambiguous. However, it does not explicitly differentiate from sibling tools, which appear to be unrelated utilities (e.g., calculators, generators, converters), so it lacks sibling differentiation.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It mentions multiple transformation modes (e.g., case-convert, slug, detect) but does not specify scenarios, prerequisites, or exclusions. Without such context, users must infer usage from the tool name and parameters alone.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
subscribe-proSubscribe to JobDoneBot ProAInspect
Create a checkout session for JobDoneBot Pro (500 JPY/month). Available plans: JobDoneBot Pro (500 JPY/month), JobDoneBot Team (2980 JPY/month). Returns a Stripe checkout URL.
| Name | Required | Description | Default |
|---|---|---|---|
| plan | No | Subscription plan | pro-monthly |
| No | Email address for the subscription (optional) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden. It discloses that the tool creates a checkout session and returns a Stripe URL, which implies a transactional, non-destructive operation that likely requires payment processing. However, it doesn't mention authentication needs, rate limits, error conditions, or what happens after the checkout (e.g., subscription activation timing).
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is efficiently structured in two sentences: the first states the core purpose and price, the second lists available plans and specifies the return value. Every word serves a purpose with zero waste.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a tool with 2 parameters, 100% schema coverage, and no output schema, the description is reasonably complete. It covers the purpose, plans, and return value. However, as a transactional tool with no annotations, it could better explain behavioral aspects like error handling or user experience flow.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already fully documents both parameters (plan with enum values and optional email). The description adds no additional parameter semantics beyond what's in the schema, such as explaining the difference between 'pro-monthly' and 'team-monthly' or when to provide the email.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the specific action ('Create a checkout session'), target resource ('JobDoneBot Pro'), and price ('500 JPY/month'). It distinguishes from sibling tools by specifying this is for subscription checkout, unlike other tools like 'check-subscription' which likely checks status.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
It provides clear context by listing available plans (Pro and Team) with their prices, indicating when to use this tool for subscription initiation. However, it doesn't explicitly state when NOT to use it or mention alternatives like direct API calls or other payment methods.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
subsidy-budget-planner補助金 収支計画メーカーBInspect
補助金申請用の収支予算書をExcel/PDFで作成 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full behavioral disclosure burden. It mentions the output format (Excel/PDF) and that it's browser-based, but doesn't address critical behavioral aspects: whether it requires authentication, if it saves data, what happens on failure, rate limits, or what the generated file contains. For a document creation tool with zero annotation coverage, this is insufficient.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise - a single sentence that efficiently communicates the core functionality. Every word earns its place: identifies the document type, purpose, output formats, and implementation method. No wasted words or redundant information.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a document generation tool with no annotations and no output schema, the description is incomplete. It doesn't explain what the generated budget statement contains, whether it includes templates or calculations, how users interact with the browser-based interface, or what happens after generation. The single sentence leaves too many operational questions unanswered.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The tool has 0 parameters with 100% schema description coverage, so there's no parameter documentation burden. The description appropriately doesn't discuss parameters since none exist. It earns a 4 because it correctly focuses on the tool's purpose rather than attempting to describe non-existent parameters.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: creating a subsidy application budget statement in Excel/PDF format. It specifies the verb ('作成' - create/make) and resource ('補助金申請用の収支予算書' - subsidy application budget statement) with output format details. However, it doesn't differentiate from siblings like 'balance-sheet' or 'cash-flow-statement' which might also create financial documents.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides minimal usage guidance - only that it's for subsidy applications. It doesn't specify when to use this versus other financial document tools (like 'invoice-generator' or 'expense-report'), nor does it mention prerequisites, alternatives, or exclusions. The context is too narrow for proper tool selection.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
take-home-pay-calcTake-Home Pay CalcCInspect
Net salary from gross with full tax and insurance breakdown.
| Name | Required | Description | Default |
|---|---|---|---|
| age | No | Age | |
| dependents | No | Number of dependents | |
| annualIncome | Yes | Annual income in JPY | |
| employmentType | No | Employment type | employee |
| spouseDependentDeduction | No | Spouse dependent deduction |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. While it mentions 'full tax and insurance breakdown,' it doesn't specify what jurisdiction's taxes apply (implied as JPY suggests Japan), whether calculations are real-time or estimates, what specific taxes/insurances are included, or how results are formatted. For a financial calculation tool with zero annotation coverage, this is insufficient.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence that gets straight to the point without unnecessary words. It's appropriately sized for a calculator tool, though it could potentially benefit from slightly more detail given the complexity of tax calculations.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a tax calculation tool with 5 parameters, no annotations, and no output schema, the description is incomplete. It doesn't explain what the output will contain beyond 'breakdown,' doesn't specify jurisdiction or calculation methodology, and provides no error handling or limitation information. The description should do more to compensate for the lack of structured output information.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The description doesn't add any parameter-specific information beyond what's already in the schema (which has 100% coverage). It mentions 'tax and insurance breakdown' which relates to the output, not the inputs. With complete schema documentation, the baseline score of 3 is appropriate since the schema already provides all parameter descriptions, defaults, and constraints.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: calculating net salary from gross income with tax and insurance breakdown. It specifies the verb ('Net salary from gross') and resource (salary calculation), but doesn't explicitly differentiate from sibling tools like 'freelance-tax-calc' or 'resident-tax-calc' that might handle similar calculations.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. There are several sibling tools with 'calc' or 'tax' in their names (freelance-tax-calc, resident-tax-calc, withholding-tax-calc, etc.) that might overlap in functionality, but the description offers no comparison or context for choosing this specific calculator.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
tax-return-calcTax Return CalcCInspect
Japan income tax and resident tax with progressive rates.
| Name | Required | Description | Default |
|---|---|---|---|
| donations | No | Donations (ふるさと納税等) in JPY | |
| blueReturn | No | Blue return deduction level | none |
| dependents | No | Number of dependents | |
| miscRevenue | No | Miscellaneous revenue (雑所得) in JPY | |
| miscExpenses | No | Miscellaneous expenses in JPY | |
| lifeInsurance | No | Life insurance premiums in JPY | |
| dividendAmount | No | Dividend income in JPY | |
| businessRevenue | No | Business revenue (事業収入) in JPY | |
| medicalExpenses | No | Medical expenses in JPY | |
| socialInsurance | No | Social insurance premiums in JPY | |
| businessExpenses | No | Business expenses (事業経費) in JPY | |
| employmentRevenue | No | Employment revenue (給与収入) in JPY |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It mentions progressive rates but does not explain key behaviors such as whether calculations are real-time, if historical data is needed, error handling, or output format. For a tax calculation tool with 12 parameters, this lack of detail is a significant gap.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence with zero wasted words. It front-loads the core purpose (tax calculation) and specifies the context (Japan, progressive rates), making it easy to parse quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a complex tax calculation tool with 12 parameters, no annotations, and no output schema, the description is insufficient. It lacks details on behavioral traits, usage context, and output format, leaving the agent with significant gaps in understanding how to effectively invoke and interpret results from this tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema fully documents all 12 parameters. The description adds no additional parameter semantics beyond implying the tool uses progressive tax rates, which is already suggested by the tool's purpose. This meets the baseline for high schema coverage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool calculates Japan income tax and resident tax with progressive rates, specifying both the action (calculate) and the resource (taxes). It distinguishes from siblings like 'freelance-tax-calc' and 'resident-tax-calc' by mentioning both income and resident taxes, but could be more explicit about its unique scope.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives like 'freelance-tax-calc' or 'resident-tax-calc'. The description does not mention prerequisites, target users, or scenarios where this tool is preferred, leaving the agent to infer usage from the tool name alone.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
telop-image-generatorテロップ画像生成BInspect
動画用テロップ画像を透過PNGで生成。フォント・色・影カスタマイズ (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations, the description carries the full burden. It discloses the output format (transparent PNG) and execution context (browser-based), but lacks critical behavioral details such as return value structure, file size limits, or whether the tool requires interactive user input.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely concise single sentence with parenthetical clarification. Information is front-loaded with the action (generate) and key attributes (transparent PNG, video use) appearing immediately.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
While the description identifies the output format (PNG), it lacks return value details necessary given the absence of an output schema. For a generative tool, not describing how the image is delivered (URL, base64, etc.) leaves a significant gap.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema contains 0 parameters, establishing a baseline of 4. The description mentions customization capabilities (font, color, shadow) which clarifies the tool's GUI functionality despite the empty parameter schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states it generates transparent PNG telop images for video with font/color/shadow customization. However, it does not explicitly differentiate from similar siblings like 'text-overlay-maker' or 'eyecatch-maker'.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance provided on when to use this tool versus alternatives (e.g., 'text-overlay-maker'). The 'Browser-based tool' note hints at execution context but does not constitute usage guidelines.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
terms-generatorTerms GeneratorBInspect
Create terms of service and privacy policies from questionnaire. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It states the tool creates documents but doesn't mention whether this requires user authentication, how the questionnaire is presented, what format the output takes, or any rate limits. For a document generation tool with zero annotation coverage, this leaves significant behavioral gaps.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is brief and front-loaded with the core functionality. The parenthetical 'Browser-based tool' adds useful context without verbosity. However, it could be slightly more structured by explicitly separating purpose from implementation details.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity (document generation), lack of annotations, and no output schema, the description is minimally adequate. It states what the tool does but lacks details on behavior, output format, or error handling. With zero parameters, the burden is lower, but more context would improve completeness for this type of tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description appropriately doesn't discuss parameters, and the mention of 'from questionnaire' provides context about how input is gathered without conflicting with the empty schema. A baseline of 4 is suitable for zero-parameter tools.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Create terms of service and privacy policies from questionnaire.' It specifies the verb ('Create'), resource ('terms of service and privacy policies'), and input source ('from questionnaire'). However, it doesn't explicitly differentiate from sibling tools like 'contract-generator' or 'nda-generator' that might also create legal documents, which prevents a perfect score.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides minimal usage guidance. It mentions the input source ('from questionnaire') but doesn't specify when to use this tool versus alternatives (e.g., 'contract-generator' for other agreements) or any prerequisites. The parenthetical note 'Browser-based tool' adds some context but doesn't address tool selection criteria.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
text-overlay-makerテキストオーバーレイAInspect
画像にテキストを重ねて配置。影・縁取り・グラデーション対応 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden. It adds valuable context by noting it is 'Browser-based' (execution environment) and lists supported styling effects. However, it omits critical behavioral details: how input is provided given the empty parameter schema (likely interactive file upload), output format, file size limits, and whether processing occurs client-side or server-side.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is exceptionally concise with two information-dense clauses. It front-loads the core function (text overlay) before detailing styling capabilities, ending with execution context in parentheses. No redundancy or filler text.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the empty input schema and lack of output schema/annotations, the tool's operational model is unclear. The description explains 'what' it does but inadequately explains 'how' to invoke it (critical for a 0-parameter tool). It should clarify the interactive browser-based workflow implied by 'Browser-based tool' and the absence of JSON parameters.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema contains zero parameters. Per the evaluation rules, zero parameters establishes a baseline score of 4. The description does not need to compensate for missing parameter documentation, though it could have clarified why no parameters exist (e.g., 'opens browser UI for configuration').
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the specific action (テキストを重ねて配置/overlay text), target resource (画像/images), and distinguishing capabilities (影・縁取り・グラデーション/shadows, borders, gradients). This effectively differentiates it from sibling tools like image-crop (no text), watermark (typically specific to copyright marking), or eyecatch-maker (specific to blog headers).
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
While the description explains functionality, it provides no guidance on when to select this tool over similar siblings like watermark, eyecatch-maker, or youtube-thumbnail-maker. It also fails to mention prerequisites (e.g., whether an image must be provided first) or that it is an interactive tool versus automated.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
text-statisticsText StatisticsBInspect
Analyze text for readability, vocabulary richness, and keyword extraction.
| Name | Required | Description | Default |
|---|---|---|---|
| text | Yes | Text to analyze |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It states what the tool does but doesn't cover critical aspects like whether it's read-only, what the output format looks like, any rate limits, or error conditions. For a tool with zero annotation coverage, this is a significant gap.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence with zero wasted words. It's front-loaded with the core purpose and lists the three analysis aspects clearly, making it easy to parse.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (text analysis with multiple metrics), no annotations, and no output schema, the description is minimally adequate. It states what the tool does but lacks details on output format, behavioral traits, or usage context, leaving gaps for the agent to navigate.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The schema description coverage is 100%, with the single parameter 'text' clearly documented in the schema. The description doesn't add any meaningful parameter details beyond what the schema provides, so it meets the baseline for high schema coverage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with specific verbs ('analyze text') and resources ('readability, vocabulary richness, and keyword extraction'), making it distinct from most siblings. However, it doesn't explicitly differentiate from tools like 'word-counter' or 'similarity-checker' that might also analyze text, so it's not a perfect 5.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It doesn't mention prerequisites, context, or exclusions, leaving the agent to infer usage from the purpose alone.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
thumbnail-ctr-predictorサムネCTR予測スコアBInspect
サムネイルの要素を分析してCTRスコアを予測。改善提案付き (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden. It adds crucial context with '(Browser-based tool)', signaling an interactive execution model necessary given the empty parameter schema. It also discloses that improvement suggestions are included. However, it fails to explain how the thumbnail is provided (file upload, URL, current context?), what the score range is, or whether results are persisted.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise—one sentence stating purpose and a parenthetical execution note. It is front-loaded with no wasted words. However, given the complexity of a zero-parameter predictive tool, it borders on underspecified rather than optimally concise.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the empty schema and lack of output schema or annotations, the description provides the minimum viable context by identifying the browser-based nature and core functionality. However, it lacks explanation of the input workflow (how the thumbnail reaches the analyzer) and expected output format beyond 'score' and 'suggestions'.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has zero parameters, which per guidelines establishes a baseline of 4. The description does not need to compensate for missing schema documentation, though it could have clarified the lack of parameters by explaining the browser-based input mechanism.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool analyzes thumbnail elements, predicts a CTR score, and includes improvement suggestions. It uses specific verbs (analyze, predict) and identifies the resource (thumbnail elements), distinguishing it from sibling creation tools like 'youtube-thumbnail-maker'.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No explicit guidance on when to use this versus alternatives like 'youtube-thumbnail-maker' or 'cvr-improvement-checker'. It doesn't clarify prerequisites (e.g., requiring an existing thumbnail vs. creating one) or when analysis is preferable to generation.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
timestamp-converterTimestamp ConverterBInspect
Convert Unix timestamps to/from human-readable dates.
| Name | Required | Description | Default |
|---|---|---|---|
| value | Yes | Unix timestamp (seconds/milliseconds) or ISO/date string | |
| direction | No | Conversion direction | toDate |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It mentions conversion behavior but lacks critical details: whether it handles timezone conversions, what formats are supported beyond 'ISO/date string', error handling for invalid inputs, or if it's a pure function without side effects. For a tool with no annotation coverage, this leaves significant behavioral gaps.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence with zero wasted words. It's front-loaded with the core purpose and uses the minimal necessary phrasing ('to/from' efficiently conveys bidirectionality). Every word earns its place.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's low complexity (simple conversion), 100% schema coverage, and no output schema, the description is minimally adequate. However, without annotations, it should ideally mention that it's a pure conversion function with no side effects. The lack of output format explanation (though no output schema exists) is a minor gap for completeness.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema fully documents both parameters. The description adds no parameter-specific information beyond what's in the schema (e.g., it doesn't elaborate on 'value' formats or 'direction' implications). Baseline 3 is appropriate when the schema does all the parameter documentation work.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the specific action ('convert') and resources ('Unix timestamps' and 'human-readable dates'), with bidirectional conversion indicated by 'to/from'. It distinguishes from the sibling 'timestamp-generator' (which likely creates timestamps rather than converting them) and 'date-calculator' (which performs date arithmetic rather than format conversion).
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It doesn't mention scenarios like handling different timestamp formats (seconds vs milliseconds), timezone considerations, or when to choose this over sibling tools like 'date-calculator' or 'timezone-converter'. No prerequisites or exclusions are stated.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
timestamp-generatorタイムスタンプ生成AInspect
開始時間+タイトルを入力→YouTube形式のタイムスタンプ出力 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden. It successfully discloses the browser-based nature (crucial for a tool with zero parameters) and the output format (YouTube style). It implies non-destructive generation behavior, though it could clarify whether the tool stores data or operates purely client-side.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely concise single-sentence structure using arrow notation (→) to show input-to-output flow. The Japanese text is dense but efficient, with the English parenthetical earning its place by clarifying the execution model. No wasted words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Adequate for a specialized tool with no output schema: it explains the return value (YouTube format timestamp) and execution model. However, given the large sibling toolset including 'timestamp-converter', the description should explicitly differentiate these tools to prevent selection errors.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
With zero parameters, the baseline is 4 per rubric. The description adds value by explaining what inputs are expected in the browser UI (start time, title), preventing confusion about the empty schema. It effectively compensates for the schema's lack of parameter definitions by describing the interactive input flow.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool generates YouTube-format timestamps from start times and titles, using specific verbs and resources. However, it loses one point for not explicitly distinguishing its role from the sibling 'timestamp-converter' (generation vs. conversion), though the 'browser-based' hint implies an interactive workflow distinct from automated conversion.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The parenthetical '(Browser-based tool)' provides critical context that this launches an interactive UI rather than accepting parameters programmatically, which informs when to use it. However, it lacks explicit guidance on when to choose this over the sibling 'timestamp-converter' or other alternatives.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
timezone-converterタイムゾーン変換CInspect
世界の主要都市の時差を即計算。サマータイム自動対応
| Name | Required | Description | Default |
|---|---|---|---|
| dateTime | Yes | Date-time string (ISO 8601 or YYYY-MM-DDTHH:mm) | |
| toTimezones | No | Target timezones | |
| fromTimezone | No | Source timezone (IANA format, e.g. Asia/Tokyo) | Asia/Tokyo |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. While it mentions automatic daylight saving time handling, it doesn't cover important aspects like error handling (e.g., invalid timezone names), rate limits, authentication needs, or what the output format looks like. For a tool with no annotation coverage, this leaves significant gaps in understanding its behavior.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise - just two short phrases that communicate the core functionality efficiently. Every word earns its place: the first phrase states the purpose, the second adds a key feature (daylight saving time handling). No wasted words or redundant information.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given that there are no annotations and no output schema, the description should provide more complete context. While concise, it doesn't explain what the tool returns, how errors are handled, or any limitations. For a tool with 3 parameters and no structured output documentation, this leaves the agent with insufficient information to use it effectively.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all three parameters thoroughly. The description doesn't add any parameter-specific information beyond what's in the schema. It implies timezone conversion but provides no additional syntax or format details. This meets the baseline expectation when schema coverage is high.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: converting time differences between major cities with automatic daylight saving time handling. It uses specific verbs ('計算' - calculate) and resources ('世界の主要都市の時差' - time differences of major world cities). However, it doesn't explicitly differentiate from sibling tools like 'timestamp-converter' or 'date-calculator', which might have overlapping functionality.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It doesn't mention sibling tools like 'timestamp-converter' or 'date-calculator', nor does it specify prerequisites or exclusions. The agent must infer usage context solely from the tool name and description.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
title-multilingualタイトル多言語変換AInspect
日本語タイトルを英語/中国語/韓国語に変換。海外展開用 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It discloses the 'Browser-based' nature, which is crucial behavioral context (suggests UI interaction rather than API processing), but omits details about input mechanisms, output format, or whether the tool waits for user interaction.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely concise at roughly 10 words plus a parenthetical. The core function is front-loaded. Minor deduction because the '(Browser-based tool)' tag, while informative, breaks the flow and raises operational questions without explanation.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a tool with zero parameters and no output schema, the description adequately covers the basic transformation intent. However, given the ambiguity of how input is received (empty schema yet requires a title) and the lack of output specification, the description leaves significant operational gaps unexplained.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema has zero parameters, which per the evaluation rules establishes a baseline score of 4. The description correctly implies no schema parameters are needed, though it doesn't explain how the Japanese title input is actually provided (likely through the browser UI mentioned).
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description clearly states the function (convert Japanese titles), target languages (English/Chinese/Korean), and intended use case (overseas expansion). It distinguishes from sibling content tools like 'rewrite-assistant' through its specific multilingual scope, though the 'Browser-based' parenthetical is slightly confusing in an MCP context.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Provides implied usage context ('海外展開用' / for overseas expansion) suggesting when to use it, but lacks explicit when-NOT-to-use guidance or comparison to alternatives like 'rewrite-assistant' or 'meta-description-generator'.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
trial-balanceTrial BalanceBInspect
Generate trial balance sheets from account balances. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It states the tool generates trial balance sheets but doesn't clarify if this is a read-only operation, what permissions might be needed, how data is sourced, or any rate limits. The 'Browser-based tool' note adds minimal context about execution but lacks critical behavioral details for a tool that likely processes financial data.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is very concise with two short phrases, front-loading the core purpose and adding a brief note about the tool's nature. There's no wasted text, but it could be slightly more informative without losing efficiency, such as by clarifying the output format or data requirements.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity of generating financial reports, the lack of annotations, no output schema, and minimal behavioral details in the description, it's incomplete. The agent needs more information about what the tool returns, how it handles errors, or any dependencies, making it inadequate for reliable use without additional context.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description doesn't add parameter details, which is appropriate here, but it could have mentioned if any implicit inputs (like account data sources) are required. Since there are no parameters, a baseline of 4 is justified as the description doesn't need to compensate for schema gaps.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with a specific verb ('Generate') and resource ('trial balance sheets from account balances'), making it easy to understand what the tool does. However, it doesn't explicitly differentiate from its sibling 'balance-sheet' tool, which appears to be a related financial reporting tool, leaving some ambiguity about when to choose one over the other.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives like 'balance-sheet' or other financial tools in the sibling list. The mention 'Browser-based tool' hints at the execution environment but doesn't offer usage context, prerequisites, or exclusions, leaving the agent without operational direction.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
unit-converter単位変換CInspect
長さ・重さ・温度・面積・容量など主要単位を即変換
| Name | Required | Description | Default |
|---|---|---|---|
| value | Yes | Value to convert | |
| toUnit | Yes | Target unit (e.g. mi, kg, f, mb) | |
| fromUnit | Yes | Source unit (e.g. km, lb, c, gb) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions '即変換' (immediate conversion), implying speed, but doesn't cover critical aspects like error handling (e.g., invalid units), precision, supported unit categories beyond the listed examples, or whether it's a read-only operation. For a tool with no annotations, this leaves significant gaps in understanding its behavior.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is very concise—a single sentence in Japanese that efficiently states the tool's function. It's front-loaded with the core purpose and includes examples of unit categories. There's no wasted text, though it could benefit from slightly more detail given the lack of annotations.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity of unit conversion (multiple categories, potential for errors) and the absence of annotations and output schema, the description is incomplete. It doesn't explain return values, error conditions, supported units beyond a vague list, or behavioral traits. For a tool with no structured support, this leaves the agent under-informed.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all three parameters (value, fromUnit, toUnit) with basic descriptions. The description adds no additional parameter semantics beyond implying support for categories like length and weight, which doesn't clarify unit formats or constraints. This meets the baseline for high schema coverage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: converting major units like length, weight, temperature, area, and volume. It specifies the action ('即変換' - immediate conversion) and resource (units), making it easy to understand. However, it doesn't explicitly differentiate from sibling tools like 'color-converter' or 'timestamp-converter', which handle different types of conversions.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives. The description lists categories of units it handles but doesn't mention when to choose it over other conversion tools (e.g., 'color-converter' for colors, 'timestamp-converter' for time). There's no context about prerequisites, limitations, or typical use cases.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
upscalerAI Image UpscalerBInspect
AI super-resolution upscaling up to 4x. Supports PNG, JPG, WebP. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It mentions 'AI super-resolution upscaling up to 4x' and 'Browser-based tool', which gives some behavioral context (AI-based, browser execution). However, it lacks critical details: whether it modifies original files, auth needs, rate limits, output format, or error handling. For a tool with zero annotation coverage, this is insufficient.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise: three short phrases packed with key information (function, formats, context). Every sentence earns its place, with no wasted words. It's front-loaded with the core purpose, making it easy to scan.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given 0 parameters and no output schema, the description is adequate but has gaps. It covers purpose and formats, but lacks details on how the tool behaves (e.g., output format, file handling), especially with no annotations. For a tool with no structured data, it provides a baseline but misses completeness for safe invocation.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameters need documentation. The description adds value by implying input formats ('Supports PNG, JPG, WebP'), which compensates for the lack of parameter schema. However, it doesn't specify how inputs are provided (e.g., file upload, URL), keeping it at 4 instead of 5.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'AI super-resolution upscaling up to 4x' specifies the verb (upscaling) and resource (images), and 'Supports PNG, JPG, WebP' adds format details. It distinguishes from siblings like 'image-resizer' or 'smart-resize' by emphasizing AI-based super-resolution. However, it doesn't explicitly differentiate from all image-related tools (e.g., 'bg-remover'), keeping it at 4 rather than 5.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides minimal usage guidance: 'Supports PNG, JPG, WebP' implies when to use it for those formats, and '(Browser-based tool)' hints at context. However, it lacks explicit when-to-use vs. alternatives (e.g., vs. 'image-resizer' for non-AI resizing), prerequisites, or exclusions. No sibling tool comparisons are made, leaving usage unclear.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
url-encoderURL EncoderCInspect
URL encode/decode with parameter parsing.
| Name | Required | Description | Default |
|---|---|---|---|
| mode | No | Encode or decode | encode |
| text | Yes | Text to encode or decode | |
| encodeMode | No | Encoding mode (component recommended) | component |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It mentions 'parameter parsing' but does not clarify what that means behaviorally (e.g., how parameters are handled, error conditions, or output format). This leaves significant gaps in understanding the tool's operation.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence with no wasted words. It is front-loaded with the core function, though it could be slightly more informative without losing conciseness.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no annotations and no output schema, the description is incomplete. It fails to explain the tool's behavior, output format, or error handling, which are critical for a tool with three parameters and no structured output documentation.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema fully documents parameters. The description adds no additional meaning beyond what the schema provides, such as explaining the practical differences between encoding modes. Baseline 3 is appropriate as the schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description 'URL encode/decode with parameter parsing' states the basic function but is vague about what 'parameter parsing' entails. It distinguishes from siblings by focusing on URL encoding/decoding, but lacks specificity about the exact transformation (e.g., percent-encoding).
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance on when to use this tool versus alternatives is provided. The description does not mention any prerequisites, typical use cases, or comparisons with other tools, leaving the agent without context for selection.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
utm-builderUTM BuilderCInspect
Build and parse campaign tracking URLs.
| Name | Required | Description | Default |
|---|---|---|---|
| url | Yes | Base URL | |
| term | No | UTM term (optional) | |
| medium | Yes | UTM medium (e.g. cpc, email) | |
| source | Yes | UTM source (e.g. google, newsletter) | |
| content | No | UTM content (optional) | |
| campaign | Yes | UTM campaign name |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It states the tool can 'build and parse' URLs but doesn't clarify whether this is read-only or mutative, what the output format is, or any error handling or constraints. This is insufficient for a tool with multiple parameters and no output schema.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise—a single sentence with no wasted words. It is front-loaded and directly states the tool's core functionality without unnecessary elaboration.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (6 parameters, no annotations, no output schema), the description is inadequate. It doesn't explain what 'build and parse' entails operationally, what the output looks like, or any behavioral nuances, leaving significant gaps for an AI agent to infer correct usage.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 100% description coverage, clearly documenting all parameters. The description adds no additional semantic context beyond implying UTM-related parameters, which is already evident from the schema. This meets the baseline for high schema coverage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with specific verbs ('build' and 'parse') and resource ('campaign tracking URLs'), making it easy to understand. However, it doesn't explicitly differentiate from sibling tools, though none appear to be direct alternatives for UTM operations.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives or in what contexts. The description lacks any mention of prerequisites, typical use cases, or exclusions, leaving the agent without operational context.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
uuid-generatorUUID GeneratorBInspect
Generate random UUIDs instantly.
| Name | Required | Description | Default |
|---|---|---|---|
| count | No | Number of UUIDs to generate | |
| format | No | Output format | standard |
| version | No | UUID version | v4 |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden for behavioral disclosure. It mentions 'instantly,' which hints at performance, but lacks details on rate limits, error conditions, or output format (e.g., whether UUIDs are returned as a list or string). For a tool with no annotations, this is insufficient to inform safe and effective use.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise—a single sentence with no wasted words. It's front-loaded with the core action ('Generate random UUIDs') and adds a useful qualifier ('instantly'). Every word earns its place, making it easy to parse quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's low complexity (generates identifiers), 100% schema coverage, and no output schema, the description is minimally adequate. However, without annotations or output details, it lacks completeness for behavioral aspects like error handling or return structure, which could be important for integration.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The schema description coverage is 100%, with clear descriptions for all parameters (count, format, version). The tool description adds no parameter-specific information beyond what the schema provides, so it meets the baseline score of 3. No compensation is needed given the comprehensive schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with a specific verb ('Generate') and resource ('UUIDs'), and the adverb 'instantly' adds useful context about speed. However, it doesn't explicitly differentiate from sibling tools like 'hash-generator' or 'timestamp-generator' that also generate identifiers, though UUIDs are distinct enough that this isn't critical.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives. The description doesn't mention prerequisites, typical use cases, or comparisons to similar tools in the sibling list (e.g., 'hash-generator' for other types of identifiers). This leaves the agent without context for tool selection.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
vector-viewerVector ViewerBInspect
Preview .ai/.eps files without Adobe Illustrator. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It states it's a 'Browser-based tool,' which adds useful context about accessibility. However, it lacks details on behavioral traits like supported file sizes, output format, whether it's read-only or modifies files, or any limitations. For a tool with zero annotation coverage, this is a significant gap.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise and front-loaded: one sentence states the core functionality, and a parenthetical adds key context. Every word earns its place with zero waste, making it easy to parse quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool has 0 parameters, no annotations, and no output schema, the description is minimally adequate. It covers the basic purpose and context but lacks details on behavior, output, or limitations. For a tool with no structured data, it should provide more completeness, but it meets a bare minimum.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The tool has 0 parameters, and schema description coverage is 100%. With no parameters, the description doesn't need to add parameter semantics. The baseline for 0 parameters is 4, as it appropriately avoids unnecessary details.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Preview .ai/.eps files without Adobe Illustrator.' It specifies the action (preview), target resources (.ai/.eps files), and key constraint (no Adobe Illustrator needed). However, it doesn't explicitly differentiate from sibling tools, as none appear to be direct alternatives for vector file previewing.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage context: when you need to preview vector files and lack Adobe Illustrator. It mentions 'Browser-based tool,' suggesting it's accessible via web. However, it doesn't provide explicit when-to-use guidance, alternatives, or exclusions compared to other tools.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
video-duration-calculator動画尺カリキュレーターBInspect
台本文字数→動画尺を推定。話速調整・尺配分計画 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It successfully signals the browser-based nature (crucial for an MCP tool), but fails to clarify what the agent receives in return (a URL? a calculation result? nothing?), whether the tool blocks waiting for user interaction, or if there are any side effects. The mention of 'speech speed adjustment' hints at configurable behavior but doesn't explain how it's controlled given the empty parameter schema.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is efficiently structured with the primary function stated first (台本文字数→動画尺を推定), followed by feature details (話速調整・尺配分計画), and execution context (Browser-based tool). Every element earns its place, though the parenthetical format of the final clause is slightly informal.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complete absence of annotations, output schema, and input parameters, the description adequately explains the conversion logic but leaves critical gaps regarding the browser interaction model and return values. For a browser-based tool, the agent needs to know whether to expect a URL, a rendered result, or synchronous completion, none of which are specified.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema contains zero properties. According to calibration rules, zero parameters establishes a baseline score of 4. The description implicitly references conceptual inputs (script word count) to explain the tool's purpose, but since the schema is empty (likely because it opens a browser UI), no parameter documentation is required from the description.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the core function (estimating video duration from script word count) using specific verbs (推定/estimate) and resources (台本文字数/script word count, 動画尺/video duration). It distinguishes from siblings like video-idea-generator or video-optimizer by focusing on duration calculation and planning (話速調整・尺配分計画). Score 4 instead of 5 because the 'Browser-based tool' qualifier introduces slight ambiguity about execution model without elaboration.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no explicit guidance on when to use this tool versus alternatives like video-optimizer or views-simulator, nor does it specify prerequisites (e.g., having a script ready). While the mention of 'speech speed adjustment' and 'duration allocation planning' implies use cases, there is no explicit 'when-to-use' or 'when-not-to-use' guidance.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
video-idea-generator動画ネタジェネレーターBInspect
ジャンル×トレンドから動画企画を量産 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It adds the parenthetical '(Browser-based tool)' indicating the execution environment, but lacks details on return format, authentication requirements, or whether the browser opens interactively.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description efficiently conveys the tool's function in a single Japanese sentence without redundancy, placing the essential scope (genre × trends → video ideas) first and adding the execution context (browser-based) parenthetically without clutter.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a zero-parameter tool without output schema or annotations, the description captures the core function and execution mode, but the lack of return value description or browser interaction details leaves gaps in the agent's understanding of invocation outcomes.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
With zero parameters in the input schema, the baseline score applies per rubric. The description's mention of 'ジャンル×トレンド' provides semantic context for the conceptual inputs driving generation despite the absence of formal parameters.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description uses the specific verb '量産' (mass produce) and identifies the resource as '動画企画' (video plans), while the mechanism 'ジャンル×トレンドから' distinguishes it from siblings like article-outline-generator or youtube-description-generator that handle different content types or output formats.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description states what the tool does but provides no guidance on when to prefer it over related content generation tools (e.g., article-outline-generator for articles vs videos) or prerequisites for invocation.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
video-optimizerVideo OptimizerBInspect
Optimize videos for Instagram Reels and Stories. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden for behavioral disclosure. It mentions 'Browser-based tool', which hints at execution environment but doesn't clarify critical traits: whether it's read-only or destructive (e.g., modifies original files), authentication needs, rate limits, output format, or error handling. For a tool that likely processes media files, this lack of detail on safety and behavior is a significant gap, making it hard for an agent to predict outcomes.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is very concise with two brief phrases: 'Optimize videos for Instagram Reels and Stories. (Browser-based tool)'. It's front-loaded with the core purpose and adds a useful environmental note. There's no wasted text, but it could be slightly more structured (e.g., separating purpose from context) for clarity, preventing a perfect score.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity of video optimization (likely involving file processing, format changes, or quality adjustments), the description is incomplete. With no annotations, no output schema, and minimal behavioral detail, it fails to provide enough context for safe and effective use. Siblings include various media tools, but this doesn't explain how it differs or what specific optimizations it performs, leaving significant gaps in understanding.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description doesn't add parameter details, which is appropriate here. It gets a baseline 4 because with zero parameters, the description's role is minimal for this dimension, and it doesn't introduce confusion or redundancy regarding inputs.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Optimize videos for Instagram Reels and Stories' with the specific verb 'optimize' and target 'videos'. It distinguishes from siblings like 'video-duration-calculator' or 'video-idea-generator' by focusing on optimization rather than analysis or ideation. However, it doesn't explicitly differentiate from tools like 'smart-resize' or 'image-resizer' that might also handle video adjustments, keeping it from a perfect score.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides minimal guidance: it implies use for Instagram content creation with 'for Instagram Reels and Stories', but offers no explicit when-to-use criteria, prerequisites, or alternatives. It doesn't specify if this is for format conversion, compression, aspect ratio adjustment, or other optimizations, nor does it compare to sibling tools like 'format-converter' or 'image-resizer' that might overlap. This leaves the agent with insufficient context for optimal tool selection.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
views-simulator再生数シミュレーターAInspect
登録者数×CTR×インプレッションから再生数を試算 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden. It successfully discloses the calculation methodology and the browser-based execution context (critical for an MCP tool as it implies UI interaction rather than JSON data return). However, it omits details about what the user receives (URL, raw calculation, or interactive form) and whether results persist.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single efficient sentence with a parenthetical qualifier. No words are wasted; the formula is front-loaded immediately, and the browser context is appended without clutter. Every element earns its place.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a zero-parameter browser tool without output schema, the description adequately covers the calculation logic and execution environment. It could be improved by explicitly stating that the tool opens a web interface or returns a URL, but the current description is sufficient for agent selection.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema has zero parameters, establishing a baseline of 4. The description adds value by identifying the three conceptual inputs required for the calculation (subscribers, CTR, impressions), preparing the user for what data the browser interface will request, even though these aren't JSON schema parameters.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool calculates estimated view counts (再生数を試算) using the specific formula 'subscribers × CTR × impressions' (登録者数×CTR×インプレッション). The 'Browser-based tool' parenthetical distinguishes its execution model from API-based calculation siblings like youtube-revenue-calculator.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
While the formula implies usage context (when you need to forecast views from subscriber base and CTR), there is no explicit guidance on when to choose this over siblings like youtube-revenue-calculator or thumbnail-ctr-predictor. The 'browser-based' hint suggests interactive use but doesn't clarify prerequisites.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
watermarkWatermarkBInspect
Add text or logo watermarks to images in batch. (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions 'Browser-based tool', which gives some context about the execution environment, but lacks details on permissions, rate limits, output format, or error handling. For a batch processing tool with zero annotation coverage, this is insufficient.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise and front-loaded: 'Add text or logo watermarks to images in batch. (Browser-based tool)'. Both sentences add value—the first states the core function, and the second provides execution context—with zero waste.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool has 0 parameters, no annotations, and no output schema, the description is minimally complete. It covers the basic purpose and execution environment but lacks details on behavioral traits (e.g., what 'batch' entails, supported formats, limitations). For a tool with no structured data, more context would be helpful.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0 parameters with 100% coverage, so no parameter documentation is needed. The description does not discuss parameters, which is appropriate. Baseline is 4 for 0 parameters, as it avoids unnecessary details.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Add text or logo watermarks to images in batch.' It specifies the action (add), resource (watermarks to images), and scope (batch). However, it does not explicitly differentiate from sibling tools like 'image-crop' or 'image-resizer', which are also image manipulation tools but for different purposes.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides minimal usage guidance. It mentions 'Browser-based tool', which implies it runs in a browser context, but does not specify when to use this tool versus alternatives (e.g., for watermarking vs. other image edits) or any prerequisites. No explicit when/when-not instructions or named alternatives are provided.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
withholding-tax-calcWithholding Tax CalcCInspect
Auto 10.21%/20.42% rate selection with reverse calculation.
| Name | Required | Description | Default |
|---|---|---|---|
| amount | Yes | Payment amount in JPY | |
| taxRate | No | Consumption tax rate (10 or 8) | |
| incomeType | No | Income type | freelance |
| taxHandling | No | Tax handling (exclusive=税抜, inclusive=税込) | exclusive |
| calcDirection | No | Calculation direction | gross-to-net |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions 'Auto 10.21%/20.42% rate selection' and 'reverse calculation,' but it doesn't explain how rates are selected (e.g., based on income type), what the output includes, or any limitations (e.g., currency, rounding). For a tax calculation tool with no annotations, this leaves critical behavioral aspects unclear.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence with zero waste. It's front-loaded with key information (auto rate selection and reverse calculation), making it appropriately sized for the tool's complexity.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (5 parameters, no output schema, no annotations), the description is inadequate. It doesn't explain the output format, how rates are applied, or any edge cases (e.g., minimum amounts, rounding rules). For a tax calculation tool, users need more context to understand the results and limitations.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all parameters thoroughly. The description adds minimal value by implying automatic rate selection (related to 'incomeType' and 'taxRate') and reverse calculation (related to 'calcDirection'), but it doesn't provide additional syntax or format details beyond what the schema specifies. Baseline 3 is appropriate when the schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description states the tool performs 'Auto 10.21%/20.42% rate selection with reverse calculation,' which indicates it calculates withholding tax with automatic rate selection and bidirectional calculation. However, it's vague about what 'reverse calculation' means (likely referring to gross-to-net or net-to-gross directions), and it doesn't clearly distinguish this from sibling tools like 'freelance-tax-calc' or 'take-home-pay-calc' in the same domain.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No explicit guidance is provided on when to use this tool versus alternatives. The description mentions automatic rate selection and reverse calculation, but it doesn't specify scenarios, prerequisites, or exclusions. Without this, users must infer usage from the tool name and parameters alone.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
word-counterWord CounterCInspect
Count characters, words, and estimate reading time.
| Name | Required | Description | Default |
|---|---|---|---|
| text | Yes | Text to analyze | |
| language | No | Text language | auto |
| countSpaces | No | Count spaces | |
| countNewlines | No | Count newlines |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions what the tool does but fails to describe key behavioral traits: it doesn't specify if this is a read-only operation, what the output format looks like (e.g., JSON structure), or any performance considerations like rate limits. For a tool with no annotations, this is a significant gap in transparency.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise and front-loaded: 'Count characters, words, and estimate reading time.' It uses a single, efficient sentence with zero waste, clearly stating the core functionality without unnecessary details. This makes it easy for an agent to quickly understand the tool's purpose.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity (a text analysis tool with 4 parameters) and lack of annotations and output schema, the description is incomplete. It doesn't explain the return values, behavioral aspects like mutability, or error handling. For a tool with no structured output information, the description should provide more context to guide the agent effectively.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, meaning the input schema already documents all parameters thoroughly. The description adds no additional semantic context beyond implying analysis of text, which is covered by the schema. This meets the baseline score of 3, as the schema handles parameter documentation adequately without extra value from the description.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Count characters, words, and estimate reading time.' It specifies the verb ('Count') and resources ('characters, words, reading time'), making the function unambiguous. However, it doesn't explicitly differentiate from sibling tools like 'text-statistics' or 'string-transform', which might offer overlapping functionality, preventing a perfect score.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It lacks context about prerequisites, scenarios where it's most useful, or comparisons to sibling tools such as 'text-statistics' or 'string-transform'. This absence of usage instructions leaves the agent without clear direction for tool selection.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
xml-json-converterXML-JSON ConverterBInspect
Convert between XML and JSON formats.
| Name | Required | Description | Default |
|---|---|---|---|
| input | Yes | XML or JSON string to convert | |
| direction | No | Conversion direction | auto |
| indentSize | No | Output indent size | |
| preserveNamespaces | No | Preserve XML namespaces |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It states the conversion action but lacks details on error handling (e.g., invalid input), performance (e.g., speed or size limits), or output characteristics (e.g., structure of converted data). For a tool with no annotations, this leaves significant gaps in understanding its behavior.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, clear sentence with zero wasted words. It is front-loaded with the core purpose and efficiently communicates the tool's function without unnecessary elaboration.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (format conversion with 4 parameters) and no output schema, the description is minimally adequate. It states what the tool does but lacks details on output format, error cases, or usage context. With no annotations and incomplete behavioral transparency, it leaves room for improvement in guiding effective use.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 100% description coverage, so parameters are well-documented in the schema itself. The description adds no additional semantic information beyond implying conversion between formats, which is already covered by the schema's parameter descriptions. This meets the baseline for high schema coverage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with the verb 'convert' and specifies the formats involved ('XML and JSON'). It distinguishes itself from siblings by focusing on format conversion, which is unique among the listed tools. However, it doesn't explicitly differentiate from potential similar tools like 'format-converter' or 'yaml-json' beyond the name.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It doesn't mention prerequisites, such as needing valid XML/JSON input, or compare it to other conversion tools like 'format-converter' or 'yaml-json'. Usage is implied by the name and purpose but not explicitly stated.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
yahoo-auction-calculatorヤフオク手数料計算機BInspect
落札手数料(10%/8.8%)+送料→利益。プレミアム会員比較 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description must carry the full behavioral burden. It successfully explains the calculation logic (落札手数料+送料→利益) and identifies the browser-based interaction mode, but omits authentication requirements, rate limits, caching behavior, or persistence details expected for a mutation-capable tool.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely compact single-sentence structure with zero redundancy. Each element serves specific purpose: fee rates (10%/8.8%), calculation flow (→利益), feature (プレミアム会員比較), and modality ((Browser-based tool)). Information density is high but readable.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
While the description covers the specific calculation domain and Yahoo Auction context adequately, it lacks description of return values or output format (no output schema exists to compensate). For a tool with zero inputs, the conceptual description suffices but the omission of what data structure or calculation results are returned leaves a gap in agent understanding.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema contains 0 parameters, which establishes a baseline score of 4 per evaluation rules. The description implies conceptual inputs (winning bid amount, shipping costs) and outputs (profit margin) without explicit parameter documentation, appropriate for an interactive parameter-free calculator.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly identifies the tool as a Yahoo Auction (ヤフオク) fee calculator, specifying the fee structure (10%/8.8%), shipping costs, and profit calculation. It distinguishes from sibling calculators (mercari-calculator, rakuten-fee-calculator) by naming platform-specific rates and premium member comparison features.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to select this tool versus similar marketplace calculators (mercalculator, yahoo-shopping-calculator, etc.). The only usage hint is '(Browser-based tool)' indicating interaction modality, but lacks decision criteria or prerequisites.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
yahoo-pr-option-simulatorYahoo! PRオプション効果シミュレーターBInspect
PRオプション料率→損益・表示順位改善効果を試算 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations, the description carries the full burden. It successfully indicates the tool is 'browser-based' (suggesting interactive web interface usage) and specifies the calculation outputs (pl and ranking effects). However, it lacks details on data persistence, authentication requirements, or side effects.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely concise with zero waste: a single Japanese sentence delivering the core value proposition followed by a parenthetical behavioral hint. Every element earns its place and is front-loaded.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Adequate for a zero-parameter simulation tool. The cover critical information (function, platform, interaction mode). However, given the complexity of Yahoo PR option mechanics (bid multipliers, impression effects), the description could benefit from mentioning expected input types (e.g., 'input your current bid rate') even if passed through the browser interface.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema contains zero parameters. Per scoring rules, this establishes a baseline score of 4. The description appropriately makes no reference to parameters since none exist.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool simulates/calculates (試算) the effects of PR option rates on profit/loss and display rankings (損益・表示順位改善効果), providing specific outputs. It distinguishes from generic siblings like yahoo-shopping-calculator by focusing specifically on PR option cost-benefit analysis.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance provided on when to use this tool versus alternatives like yahoo-shopping-calculator or rakuten-rpp-calculator. The 'browser-based' note hints at the interaction mode but doesn't clarify selection criteria.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
yahoo-shopping-calculatorYahoo!ショッピング手数料計算BInspect
ストアポイント+決済+PRオプション→利益計算 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden. It discloses the calculation scope (points + payment + PR → profit) and adds the '(Browser-based tool)' notation, hinting at execution environment, but fails to clarify what 'browser-based' entails (e.g., opens UI vs. returns JSON) or whether side effects occur.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely efficient at 11 words. The formula-like structure 'A+B+C→Result' front-loads the value proposition. The parenthetical '(Browser-based tool)' is the only extraneous element but does not significantly bloat the description.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given zero input parameters and no output schema, the description should explain the return value or side effects (e.g., 'returns profit margin' or 'opens browser'). It explains the calculation logic but omits what the tool actually returns or does upon invocation, leaving a gap for an agent trying to integrate the result into a workflow.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
With zero parameters, baseline is 4. The description adds semantic value by enumerating the conceptual inputs (store points, payment, PR options) that drive the calculation, helping the agent understand the tool's purpose despite the empty schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool calculates profit (利益計算) from specific inputs: store points (ストアポイント), payment fees (決済), and PR options. These domain-specific terms distinguish it from sibling calculators like amazon-fba-calculator or mercari-calculator. However, it relies on the title 'Yahoo!ショッピング' for platform context rather than stating it explicitly in the description body.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
There is no guidance on when to use this specific calculator versus similar siblings like rakuten-fee-calculator or shopify-profit-calculator. No prerequisites (e.g., requiring Yahoo Store registration) or exclusion criteria are mentioned.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
yaml-jsonYAML-JSON ConverterBInspect
Convert between YAML and JSON formats.
| Name | Required | Description | Default |
|---|---|---|---|
| input | Yes | YAML or JSON string to convert | |
| direction | No | Conversion direction | auto |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden for behavioral disclosure. 'Convert between YAML and JSON formats' implies a read-only transformation, but doesn't specify error handling, performance characteristics, or whether the conversion preserves data fidelity. For a tool with no annotation coverage, this leaves significant behavioral questions unanswered.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise at just 5 words: 'Convert between YAML and JSON formats.' Every word earns its place - it specifies the action, the source, and the target formats with zero waste. It's front-loaded with the core purpose.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's simple purpose (format conversion), 2 parameters with 100% schema coverage, and no output schema, the description is minimally adequate. However, without annotations and with many similar sibling tools, it lacks context about when this specific conversion is needed versus other format tools. The description covers the basic 'what' but not the 'when' or 'why'.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents both parameters thoroughly. The description doesn't add any parameter semantics beyond what's in the schema - it doesn't explain what 'auto' direction means, provide examples of valid input strings, or clarify edge cases. Baseline 3 is appropriate when the schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: 'Convert between YAML and JSON formats.' It specifies the verb ('convert') and resources/formats involved. However, it doesn't distinguish this tool from its sibling 'xml-json-converter' or other format converters in the list, which would require explicit differentiation to earn a 5.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. With many sibling tools including 'xml-json-converter', 'json-to-csv', 'format-converter', and 'base64-converter', there's no indication of when YAML-JSON conversion is specifically needed or when other format tools might be more appropriate.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
year-end-adj-calcYear-End AdjustmentCInspect
Calculate refund or additional tax owed.
| Name | Required | Description | Default |
|---|---|---|---|
| hasSpouse | No | Has spouse | |
| annualBonus | No | Annual bonus in JPY | |
| withheldTax | No | Tax already withheld in JPY | |
| annualSalary | Yes | Annual salary in JPY | |
| spouseIncome | No | Spouse income in JPY | |
| hasHousingLoan | No | Has housing loan | |
| socialInsurance | No | Social insurance premiums in JPY | |
| housingLoanBalance | No | Housing loan balance in JPY | |
| earthquakeInsurance | No | Earthquake insurance premium in JPY |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It states the tool calculates refunds or additional tax owed, implying a read-only computation without side effects, but it doesn't specify whether this is for a specific tax type (e.g., income tax), jurisdiction (e.g., Japan, inferred from JPY currency), or year. It also omits details like error handling, rate limits, or authentication needs, which are critical for a tool with 9 parameters.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence with zero waste—'Calculate refund or additional tax owed'—that directly states the tool's function. It's appropriately sized and front-loaded, making it easy to parse without unnecessary elaboration.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (9 parameters, no annotations, no output schema), the description is incomplete. It doesn't explain the calculation logic, output format (e.g., whether it returns a number or structured data), or context (e.g., tax type or jurisdiction). For a tool with many inputs and no structured output, more detail is needed to guide effective use.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, with each parameter well-documented in the schema (e.g., 'annual salary in JPY'). The description adds no additional meaning beyond the schema, as it doesn't explain how parameters interact (e.g., how spouse income affects the calculation) or provide examples. With high schema coverage, the baseline score is 3, reflecting adequate but not enhanced parameter understanding.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description 'Calculate refund or additional tax owed' clearly states the tool's purpose with a specific verb ('calculate') and outcome, but it's vague about what exactly is being calculated (e.g., year-end tax adjustment for income tax) and doesn't distinguish it from sibling tools like 'freelance-tax-calc', 'furusato-tax-calc', 'resident-tax-calc', or 'withholding-tax-calc', which are all tax-related calculators on the same server.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus the other tax-related sibling tools (e.g., 'freelance-tax-calc', 'withholding-tax-calc'), nor does it mention any prerequisites, context, or exclusions. It lacks explicit usage instructions, leaving the agent to infer based on the tool name and parameters alone.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
youtube-description-generatorYouTube説明文ジェネレーターAInspect
タイトル+KWから最適化された説明文を自動生成 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description carries full burden. It adds 'Browser-based tool' hinting at execution environment, and discloses inputs/outputs. However, it omits safety profile (read/write nature), data persistence, rate limits, and what 'optimized' specifically means (SEO, character count, etc.).
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely concise two-part structure: Japanese function statement followed by execution context parenthetical. Every element earns its place; no redundancy or generic filler.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given zero parameters and simple generation purpose, the description covers essential invocation context (inputs needed, browser-based nature). Minor gap: undefined 'optimized' criteria. Acceptable for low-complexity tool without output schema.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema is empty (0 parameters), which per rules establishes a baseline of 4. The description compensates by specifying expected inputs (title and keywords/KW) that the schema omits, adding necessary semantic context for invocation.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description clearly states the tool 'generates optimized descriptions' (specific verb + resource) from 'title + keywords' (inputs). The title 'YouTube説明文ジェネレーター' combined with description distinguishes it from siblings like youtube-tag-generator and youtube-thumbnail-maker.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage context by specifying required inputs (title+KW), but lacks explicit guidelines on when to use this vs alternatives (e.g., manual writing) or prerequisites. It does not state when NOT to use it.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
youtube-revenue-calculatorYouTube収益計算機BInspect
再生数×RPM→広告収益+案件収益を試算。月収シミュレーション (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It adds valuable context by declaring this is a 'Browser-based tool' (implying interactive UI rather than API-style parameters) and explains the calculation scope includes both ads and sponsorships. However, it lacks information on data persistence, authentication requirements, or whether calculations are client-side only.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise and front-loaded with the calculation formula (再生数×RPM→). Every element serves a purpose: the equation explains the methodology, '試算' indicates estimation behavior, and the parenthetical clarifies the execution environment. Zero redundancy.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's simplicity (zero parameters, no annotations, no output schema), the description adequately explains the calculation logic but leaves gaps regarding the output format and specific browser interaction model. For a calculation tool without structured output documentation, mentioning what values are returned would improve completeness.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has zero parameters, which establishes a baseline of 4 per the rubric. The description adds semantic value by indicating the conceptual inputs required (再生数/views and RPM), helping the agent understand what data the tool expects despite the empty parameter schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool calculates ad revenue plus sponsorship revenue (案件収益) using the formula 'views × RPM' and performs monthly income simulation. It distinguishes itself from sibling revenue calculators by using YouTube-specific terminology (RPM) and mentioning sponsorship revenue, though it does not explicitly differentiate from affiliate-revenue-calc.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides only functional information ('calculate revenue') without stating when to use this specific tool versus alternatives like affiliate-revenue-calc or mercari-calculator. No prerequisites, constraints, or workflow guidance is provided.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
youtube-tag-generatorYouTubeタグジェネレーターBInspect
キーワードから関連タグを自動展開。500文字制限内で最適化 (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Without annotations, the description carries the full burden. It discloses the 500-character optimization constraint and crucially notes it is a 'Browser-based tool,' which implies potential UI interaction requirements. However, it lacks disclosure of output format, error handling behavior, or whether it performs external API calls versus local processing.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is compact with two efficient statements. The core action is front-loaded, followed by constraints. The parenthetical '(Browser-based tool)' is slightly awkward placement and could be clearer, but overall information density is high with minimal waste.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the simple schema (0 parameters) and lack of annotations or output schema, the description covers the essential operational context (character limit, browser-based nature). However, it fails to resolve the ambiguity regarding keyword input methodology and does not describe the expected return value structure.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
With zero parameters in the schema, the baseline score is 4 per the rubric. However, there is tension between the description referencing 'from keywords' (implying input) and the empty parameter schema. The description does not clarify how keywords are provided (possibly through the browser interaction mentioned).
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool automatically expands related tags from keywords and specifically mentions the 500-character limit constraint relevant to YouTube. While the function is clear, it doesn't explicitly differentiate from sibling tools like youtube-description-generator, relying on the tool name for that distinction.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
There is no guidance on when to use this tool versus alternatives (e.g., when to use this instead of youtube-description-generator for metadata optimization). The description only states what the tool does, not when it is appropriate or prerequisites for use.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
youtube-thumbnail-makerYouTubeサムネメーカーBInspect
YouTube用サムネイルをテンプレから即生成。CTR重視デザイン (Browser-based tool)
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden. It successfully communicates the browser-based nature (implying interactive UI), but lacks critical operational details: return value format, file persistence, authentication requirements, or whether the tool blocks execution pending user interaction.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely efficient: one sentence delivering purpose, method (templates), quality attribute (CTR-focused), and operational context (browser-based). No redundant words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given zero parameters and no output schema, the description covers the essentials but remains incomplete regarding the tool's output behavior. For a browser-based tool with no structured return, it should clarify what the agent receives (e.g., image URL, file path, or just confirmation of launch).
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema has zero parameters, warranting the baseline score of 4. The description appropriately focuses on functional behavior rather than inventing parameter documentation.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states it generates YouTube thumbnails instantly from templates with CTR-focused design. It implicitly distinguishes from generic image tools via the 'YouTube-specific' and 'Browser-based' qualifiers, though it doesn't explicitly differentiate from siblings like 'eyecatch-maker'.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No explicit guidance on when to use this tool versus alternatives like 'eyecatch-maker', 'text-overlay-maker', or 'thumbnail-ctr-predictor'. The '(Browser-based tool)' parenthetical hints at operational mode but doesn't guide selection criteria.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
Claim this connector by publishing a /.well-known/glama.json file on your server's domain with the following structure:
{
"$schema": "https://glama.ai/mcp/schemas/connector.json",
"maintainers": [{ "email": "your-email@example.com" }]
}The email address must match the email associated with your Glama account. Once published, Glama will automatically detect and verify the file within a few minutes.
Control your server's listing on Glama, including description and metadata
Access analytics and receive server usage reports
Get monitoring and health status updates for your server
Feature your server to boost visibility and reach more users
For users:
Full audit trail – every tool call is logged with inputs and outputs for compliance and debugging
Granular tool control – enable or disable individual tools per connector to limit what your AI agents can do
Centralized credential management – store and rotate API keys and OAuth tokens in one place
Change alerts – get notified when a connector changes its schema, adds or removes tools, or updates tool definitions, so nothing breaks silently
For server owners:
Proven adoption – public usage metrics on your listing show real-world traction and build trust with prospective users
Tool-level analytics – see which tools are being used most, helping you prioritize development and documentation
Direct user feedback – users can report issues and suggest improvements through the listing, giving you a channel you would not have otherwise
The connector status is unhealthy when Glama is unable to successfully connect to the server. This can happen for several reasons:
The server is experiencing an outage
The URL of the server is wrong
Credentials required to access the server are missing or invalid
If you are the owner of this MCP connector and would like to make modifications to the listing, including providing test credentials for accessing the server, please contact support@glama.ai.
Discussions
No comments yet. Be the first to start the discussion!
social-insurance-calcSocial Insurance CalcCInspectHealth, pension, and employment insurance from standard monthly income.
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It mentions 'calculates' but doesn't disclose behavioral traits such as whether it's a read-only calculation, if it requires authentication, rate limits, or what the output format is (only input is described). For a calculation tool with no annotations, this leaves significant gaps in understanding how it behaves.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, efficient sentence that front-loads the core purpose. It avoids unnecessary words, but could be slightly more specific (e.g., 'Calculates social insurance premiums for health, pension, and employment based on monthly income in Japan').
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no annotations and no output schema, the description is incomplete. It doesn't explain what the tool returns (e.g., breakdown of premiums, total deductions), which is critical for a calculation tool. The input schema is well-documented, but the overall context lacks output details and behavioral transparency.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all parameters (monthlySalary, age, industry, prefecture) with descriptions. The description adds no additional meaning beyond implying these inputs are used for insurance calculations, which is minimal value. Baseline is 3 when schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description states the tool calculates 'Health, pension, and employment insurance from standard monthly income,' which gives a general purpose but lacks specificity about what it calculates (e.g., premiums, deductions, totals). It distinguishes from siblings like 'freelance-tax-calc' or 'take-home-pay-calc' by focusing on social insurance, but doesn't clearly differentiate from similar tools like 'nhi-calc' (national health insurance) or 'resident-tax-calc'.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No explicit guidance on when to use this tool versus alternatives is provided. The description implies it's for calculating social insurance based on salary, but doesn't mention prerequisites (e.g., for employees in Japan), when not to use it (e.g., for self-employed individuals), or refer to sibling tools like 'nhi-calc' or 'withholding-tax-calc' for related calculations.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.