Skip to main content
Glama

Server Details

Official MCP server for HireSquire. Automate resume screening, candidate ranking, and interview scheduling for autonomous agents.

Status
Healthy
Last Tested
Transport
Streamable HTTP
URL

Glama MCP Gateway

Connect through Glama MCP Gateway for full control over tool access and complete visibility into every call.

MCP client
Glama
MCP server

Full call logging

Every tool call is logged with complete inputs and outputs, so you can debug issues and audit what your agents are doing.

Tool access control

Enable or disable individual tools per connector, so you decide what your agents can and cannot do.

Managed credentials

Glama handles OAuth flows, token storage, and automatic rotation, so credentials never expire on your clients.

Usage analytics

See which tools your agents call, how often, and when, so you can understand usage patterns and catch anomalies.

100% free. Your data is private.
Tool DescriptionsB

Average 3.4/5 across 24 of 24 tools scored. Lowest: 2.4/5.

Server CoherenceA
Disambiguation5/5

Each tool has a clearly distinct purpose, covering separate aspects like screening, email generation, calendar management, credits, and API info. No overlapping functionalities.

Naming Consistency4/5

Most tools follow a consistent verb_noun pattern in snake_case (e.g., create_screening, get_credit_balance). The exception is 'whoami', which is a common but nonstandard verb. Overall highly consistent.

Tool Count4/5

24 tools is slightly above the typical sweet spot (3-15) but just below the 25+ threshold. The breadth of the domain (screening, credits, calendar, email, webhooks) justifies the count, though it feels a bit heavy.

Completeness4/5

The tool set covers core workflows: screening creation/monitoring/results, email generation, calendar integration, credit management, and configuration. Minor gaps like missing update/delete for some entities, but overall sufficient for the ecosystem's purpose.

Available Tools

24 tools
bulk_generate_emailsBInspect

Generate personalized emails for multiple candidates at once. Efficient for high-volume outreach after screening.

ParametersJSON Schema
NameRequiredDescriptionDefault
toneNoEmail tone
typeYesEmail type
job_idYesThe job ID
candidate_idsYesArray of candidate IDs (max 50)
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description must carry the full burden. It only states efficiency and high-volume context, omitting critical behavioral details such as whether emails are drafts or sent, idempotency, rate limits, or error handling for invalid candidate IDs.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is very concise with two sentences, no redundancy. It is front-loaded with the main action. However, it could slightly expand on behavioral context without losing conciseness.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given no output schema, the description should explain what the tool returns (e.g., generated email texts, success status). It lacks prerequisites, error handling, and post-conditions, making it incomplete for an AI agent to use confidently.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Input schema coverage is 100%, so the baseline is 3. The description adds no additional parameter-level information beyond what the schema already provides, such as the meaning of 'tone' choices or the impact of 'type'.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states it generates personalized emails for multiple candidates, with 'bulk' in the name implying batch processing. However, it does not explicitly differentiate from the sibling tool 'generate_email' (singular), which could cause confusion.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description mentions 'high-volume outreach after screening' implying a specific use case, but it does not provide explicit when-to-use or when-not-to-use guidance, nor does it mention alternatives like 'generate_email' for single candidates.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

cancel_screening_jobBInspect

Cancel a running screening job

ParametersJSON Schema
NameRequiredDescriptionDefault
job_idYesThe job ID to cancel
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description carries the full burden. It only says 'running' but does not disclose behavior for non-running jobs, side effects, permissions, or idempotency.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, short sentence that is concise and front-loaded. Every word is necessary and contributes to the purpose.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the absence of annotations and output schema, the description lacks completeness. It does not explain error handling, return values, or constraints (e.g., if the job is already cancelled).

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The schema description for job_id is 'The job ID to cancel,' which is clear. The tool description adds no additional meaning beyond what the schema already provides. Baseline of 3.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description uses the verb 'Cancel' and the resource 'screening job', clearly indicating the action and target. It distinguishes itself from sibling tools like create_screening and get_screening_status.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides no guidance on when to use this tool instead of alternatives. It does not mention prerequisites, such as the job must be running, nor does it exclude cases like completed jobs.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

compare_candidatesAInspect

Compare multiple candidates side-by-side to evaluate their relative strengths. Includes an AI-powered Delta Analysis explaining why one candidate ranks higher and highlighting specific edges.

ParametersJSON Schema
NameRequiredDescriptionDefault
job_idYesThe unique ID of the job posting.
candidate_idsYesA list of candidate IDs to compare (minimum 2, maximum 10).
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description carries the full burden. It mentions AI-powered analysis, but does not disclose read-only nature, auth needs, or side effects. Moderate transparency.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two sentences with no waste. Front-loaded with the core action and immediately adds the key differentiator (Delta Analysis). Efficient and well-structured.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

No output schema, but description covers main outputs (Delta Analysis, ranking). Could specify output format or return structure, but for a simple read tool it's adequate.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100% with clear descriptions for both parameters. The description adds no additional parameter meaning beyond repeating the AI analysis feature. Baseline 3 is appropriate.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool's purpose: compare candidates side-by-side to evaluate relative strengths, with AI-powered Delta Analysis. It distinguishes from sibling tools as the only comparison tool.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage when needing to compare candidates, but lacks explicit when-not-to-use or alternative tools. Since no other sibling tool offers comparison, the context is clear.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

create_calendar_connectionBInspect

Connect a calendar provider (calendly, calcom) to HireSquire

ParametersJSON Schema
NameRequiredDescriptionDefault
api_keyYesAPI key for the provider
providerYesCalendar provider
calendar_idNoOptional specific calendar ID
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, and the description does not disclose behavioral traits such as side effects (e.g., overwriting an existing connection), authentication requirements, or rate limits. It simply states the action without any safety or impact notes.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, front-loaded sentence that efficiently communicates the core purpose. It is concise with no fluff, though it could benefit from a brief mention of the return value.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Without annotations or output schema, the description should compensate. It lacks details on return value (e.g., connection ID), error scenarios, or behavior when a duplicate connection exists. For a 3-parameter tool, this is incomplete.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Input schema covers 100% of parameters with descriptions. The description adds no new meaning beyond what the schema already provides (e.g., 'API key' and 'Calendar provider' are already in the schema). Baseline of 3 is appropriate.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the verb 'Connect', the resource 'calendar provider', and specifies the supported providers 'calendly, calcom' and the system 'HireSquire'. This leaves no ambiguity about the tool's function.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies the tool is for connecting a new provider but does not explicitly state when to use it versus alternatives like 'list_calendar_connections' or mention prerequisites (e.g., existing account). No guidance on when not to use it is provided.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

create_checkout_sessionAInspect

Create a Stripe checkout session for credit purchase (generates a payment link)

ParametersJSON Schema
NameRequiredDescriptionDefault
packNoCredit pack to purchase
amountNoCustom amount to purchase (min $5)
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Discloses that it generates a payment link, but lacks details on side effects (e.g., no mention of idempotency or Stripe account requirements). Annotations absent, so description carries burden.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Single sentence with no superfluous information. Efficient and front-loaded.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Minimal but sufficient for a simple tool. Lacks output details and relationship to sibling tools like purchase_credits.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100%, so description adds no extra meaning. Parameters already described in schema adequately.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Clearly states the action (create), resource (Stripe checkout session), and purpose (credit purchase, generates payment link). Distinguishes from siblings like purchase_credits.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance on when to use this tool versus alternatives (e.g., purchase_credits). Missing context about prerequisites or exclusions.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

create_interviewBInspect

Create an interview with calendar event and meeting link

ParametersJSON Schema
NameRequiredDescriptionDefault
job_idYesJob posting ID
providerNoCalendar provider (Calendly or Cal.com)
candidate_idYesCandidate ID
scheduled_atYesISO 8601 datetime
duration_minutesNoDuration in minutes (default 60)
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Without annotations, the description partially discloses behavior by mentioning calendar event and meeting link creation, but it lacks details on side effects, permissions, or data persistence. It is adequate but not thorough.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single concise sentence, efficient but could benefit from additional structure or bullet points for readability.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

The tool has no output schema, and the description does not explain what the tool returns. It covers the action but not the response format or side effects, leaving some gap for the agent.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The input schema covers all 5 parameters with descriptions. The description adds no additional parameter meaning beyond the schema, which already provides sufficient detail. Baseline 3 applies due to high schema coverage.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the verb 'create' and the resource 'interview', and specifies that it also creates a calendar event and meeting link. It is distinct from sibling tools like 'create_screening' and 'generate_meeting_link'.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance is provided on when to use this tool versus alternatives, nor any prerequisites or exclusions. For a tool that creates multiple entities, usage context would be helpful.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

create_screeningAInspect

Submit candidate resumes for AI screening against a job description. Supports raw text, base64, or file URLs. High-accuracy scoring (1-100) based on role requirements.

ParametersJSON Schema
NameRequiredDescriptionDefault
titleYesThe official job title for the position (e.g., "Senior Software Engineer").
resumesYesA list of candidate resumes to be screened.
descriptionYesThe complete job description including requirements, responsibilities, and qualifications (minimum 50 characters).
webhook_urlNoOptional HTTPS URL where a POST request will be sent upon job completion.
leniency_levelNoStrictness of the AI screening on a scale of 1-10. 1 is very lenient, 10 is extremely strict. Default is 5.
webhook_conditionsNoOptional triggers for the webhook (e.g., only trigger for candidates with score > 80).

Output Schema

ParametersJSON Schema
NameRequiredDescription
job_idNo
statusNo
messageNo
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations exist, so the description carries full burden. It discloses scoring range and input formats but does not mention whether the job is async, any rate limits, or what happens in case of errors. The note on high-accuracy scoring is helpful but insufficiently detailed.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is three concise sentences, front-loaded with the key action and then additional details. No unnecessary words or redundancy.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

With a rich input schema (100% covered) and an output schema present, the description covers the essentials. However, it could note that screening is asynchronous and that users should check status via 'get_screening_status'. The sibling tools are not referenced.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100%; description adds meaningful context: 'Supports raw text, base64, or file URLs' and 'High-accuracy scoring (1-100)' which goes beyond the schema definitions. However, it does not elaborate on the 'webhook_conditions' parameter or the scoring methodology.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description uses a specific verb ('Submit') and resource ('resumes for AI screening against a job description'), clearly distinguishing it from siblings like 'get_results' or 'cancel_screening_job'. It also lists supported input formats and scoring details.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description mentions input formats and scoring range but provides no explicit guidance on when to use this tool vs alternatives (e.g., 'compare_candidates') or when not to use it. The context is implied but not stated.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

disable_auto_reloadBInspect

Disable automatic credit reloading

ParametersJSON Schema
NameRequiredDescriptionDefault

No parameters

Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description bears full burden. It only states the action without disclosing side effects, required permissions, or impact on credit reloading behavior. Minimal transparency.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

A single, clear sentence that is appropriately concise for a simple action. No wasted words.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's simplicity (0 parameters, no output schema), the description is mostly complete. However, it could mention that it requires authentication or confirm it has no side effects, but overall it communicates the core function.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

No parameters exist, so the baseline is 4. The description does not need to add meaning for parameters, and it correctly avoids mentioning any.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the action (Disable) and resource (automatic credit reloading). The purpose is evident, and the presence of a sibling tool 'enable_auto_reload' implies this is the inverse, but no explicit differentiation is provided.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance on when to use this tool versus alternatives (e.g., enabling auto reload or other credit management tools). The description lacks context about prerequisites or scenarios.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

enable_auto_reloadBInspect

Enable automatic credit reloading when balance drops below threshold

ParametersJSON Schema
NameRequiredDescriptionDefault
amountYesAmount to reload each time (in dollars, min $5)
thresholdYesMinimum balance to trigger reload (in dollars)
payment_method_idYesStripe payment method ID to use
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations present; description carries the burden. It states 'enable' implying a state change, but does not disclose if it's idempotent, whether it overrides existing settings, or any side effects. Adequate but minimal.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Single sentence, direct, no extraneous information. Efficiently communicates the core function.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

No output schema or annotations. Description is too brief; lacks details on error states, confirmation, or interaction with other credit tools. For a tool with 3 required parameters, more context is needed.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100% with clear parameter descriptions. The tool description does not add additional meaning beyond the schema, so baseline 3 is appropriate.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Description clearly states action (enable) and resource (automatic credit reloading) with a specific condition (when balance drops below threshold). It distinguishes from sibling tools like 'disable_auto_reload' and 'purchase_credits'.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No explicit guidance on when to use this tool versus alternatives like 'disable_auto_reload' or 'purchase_credits'. Lacks prerequisites or exclusions.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

estimate_screening_costCInspect

Estimate cost for screening a number of candidates

ParametersJSON Schema
NameRequiredDescriptionDefault
candidate_countYesNumber of candidates to screen
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description must disclose behavioral traits, but it only says 'Estimate cost' without mentioning whether it is read-only, consumes credits, or has side effects.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness3/5

Is the description appropriately sized, front-loaded, and free of redundancy?

A single sentence is concise, but it lacks structure and may be too brief to convey necessary information.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's complexity (one parameter, no output schema, no annotations), the description is incomplete. It does not explain the output format, cost basis, or assumptions.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The schema covers 100% of parameters with a description for candidate_count. The description adds no additional meaning beyond what the schema provides.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the verb 'Estimate' and the resource 'cost for screening a number of candidates'. It is distinct from sibling tools that perform screening or other operations.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance on when to use this tool versus alternatives (e.g., create_screening). No prerequisites or context for when the estimate is appropriate.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

generate_emailBInspect

Generate professional candidate outreach or follow-up emails based on screening results.

ParametersJSON Schema
NameRequiredDescriptionDefault
typeYesType of email to generate.invite
job_idYesThe job ID.
candidate_idYesThe candidate ID.
custom_messageNoOptional custom message

Output Schema

ParametersJSON Schema
NameRequiredDescription
bodyNo
subjectNo
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

The description lacks behavioral context beyond generating emails. It does not disclose whether the tool sends the email, stores it, or requires permissions. Since no annotations are provided, the description carries the full burden, but it only states the action without side effects or outcomes.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single concise sentence that covers the core purpose. It is front-loaded and efficient, though it could include a brief statement about output or usage without losing conciseness.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given that there is an output schema and sibling tools, the description is insufficient. It does not describe the output format, differentiate from 'bulk_generate_emails' beyond the singular nature, or mention prerequisites like screening results being available. The description feels incomplete for a tool with multiple parameters and context signals.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, so the schema already documents all parameters. The description adds minimal meaning beyond 'based on screening results', which provides context but does not explain individual parameters beyond what the schema says.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool generates professional candidate outreach or follow-up emails based on screening results. It identifies the specific verb-resource combination and implies a single email rather than bulk, distinguishing it from the sibling tool 'bulk_generate_emails'.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description mentions 'based on screening results' which implies context after screening, but it does not explicitly state when to use this tool versus alternatives like 'bulk_generate_emails' or when not to use it. The guidance is implied rather than explicit.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_api_schemaAInspect

Get the full OpenAPI 3.0 specification for the HireSquire REST API

ParametersJSON Schema
NameRequiredDescriptionDefault

No parameters

Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description carries full burden. It states the tool 'gets' the specification, implying a read-only operation, but does not explicitly disclose non-obvious behaviors such as authentication requirements or output format.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single sentence that delivers complete information without extraneous words. It is front-loaded and efficient.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given no parameters and no output schema, the description is fairly complete. It identifies the resource and standard format (OpenAPI 3.0). However, it could improve by mentioning the output format (e.g., JSON or YAML).

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

There are no parameters, so schema coverage is 100%. The description adds value by specifying the scope ('for the HireSquire REST API'), exceeding the baseline expectation for zero-param tools.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool retrieves the full OpenAPI 3.0 specification for the HireSquire REST API. This verb+resource specificity distinguishes it from sibling tools like bulk_generate_emails or whoami.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description does not provide explicit guidance on when to use this tool versus alternatives. However, the purpose is self-evident given the clear resource it accesses, so a minimal score is assigned.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_available_slotsCInspect

Get available time slots from a calendar provider

ParametersJSON Schema
NameRequiredDescriptionDefault
dateYesDate in Y-m-d format
durationNoDuration in minutes (default 60)
providerYesCalendar provider (Calendly or Cal.com)
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are present, so the description carries full burden. It does not disclose behavioral traits like authentication requirements, rate limits, or what happens when no slots are available. The description is too brief.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single sentence that is concise and front-loaded. It could benefit from a bit more detail, but it is not wasteful.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool has 3 parameters, no output schema, and no annotations, the description should provide more context about return values, pagination, or error conditions. It lacks this information, making it incomplete for an agent.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The input schema has 100% description coverage for all three parameters, so the schema already documents them well. The description adds no additional context beyond what is in the schema, meeting the baseline.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool gets available time slots from a calendar provider, specifying a verb and resource. However, it does not differentiate from sibling tools like 'generate_meeting_link' that might also involve time slots.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance is provided on when to use this tool versus alternatives such as 'generate_meeting_link' or 'create_calendar_connection'. The description lacks context about prerequisites or typical use cases.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_credit_balanceAInspect

Get current credit balance and check if you can afford screening

ParametersJSON Schema
NameRequiredDescriptionDefault

No parameters

Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description clearly indicates a read-only operation and the core behavior. However, it omits details like response format or potential edge cases (e.g., insufficient balance), which would be expected for full transparency.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, well-crafted sentence that conveys the purpose without any superfluous words.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

The description adequately explains the tool's purpose for a simple operation, but lacks details about the return value, which would be helpful since no output schema is provided.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

No parameters exist, and the schema coverage is 100%, so the description does not need to add parameter info. Per guidelines, baseline 4 applies.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the verb 'Get' and the resource 'current credit balance', and adds a secondary purpose 'check if you can afford screening', which distinguishes it from sibling tools like purchase_credits or get_credit_transactions.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage for checking available credit before screening, but does not explicitly state when not to use it or provide alternatives among siblings.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_credit_transactionsCInspect

Get credit transaction history

ParametersJSON Schema
NameRequiredDescriptionDefault
limitNoNumber of transactions to return (default 20)
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description carries full burden. It does not disclose behavioral traits such as being read-only, pagination behavior, or rate limits. Only the existence of a limit parameter hints at behavior, but it is not explained.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is very concise (one sentence, no waste). It is front-loaded with the action. However, it may be too sparse, missing context that could be added without verbosity.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool has no output schema and no annotations, the description should provide more context about the return data, sorting, or typical use cases. It is not complete enough for an agent to fully understand the tool's behavior.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100% for the single parameter 'limit', which has a clear description. The tool description adds no extra meaning beyond the schema, meeting the baseline of 3.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description 'Get credit transaction history' clearly states the verb and resource, distinguishing it from siblings like get_credit_balance (which shows current balance) and purchase_credits. It is specific enough for basic understanding.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance on when to use this tool versus alternatives. For example, it does not mention that this shows historical transactions, while get_credit_balance shows the current balance. The agent is left to infer usage.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_resultsAInspect

Retrieve detailed screening results including candidate scores, AI-generated reasoning, and interview questions. Only available for "completed" jobs.

ParametersJSON Schema
NameRequiredDescriptionDefault
job_idYesThe unique ID of the screening job.
min_scoreNoMinimum score threshold (0-100). Only candidates with scores equal to or higher than this value will be returned.
only_top_nNoLimit the results to the top N candidates based on their score.

Output Schema

ParametersJSON Schema
NameRequiredDescription
candidatesNo
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description carries the full burden. It correctly implies a read-only operation ('Retrieve'), but does not disclose potential side-effects, authentication requirements, rate limits, or error handling. The description is adequate but not rich in behavioral detail.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is concise (two sentences) and front-loaded with the core purpose and content, followed by a key constraint. Every word adds value, with no wasted information.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

The tool has an output schema (not shown in input but flagged), so description does not need to explain return values. It covers the essential: what data is returned and when it is available. Given the simple parameter set (3 params, 1 required) and presence of a constraint, the description is nearly complete. Lacks only a mention of related siblings for optimal guidance.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100%, and all parameters are already well-described in the input schema (job_id, min_score, only_top_n). The description does not add any additional meaning or usage context for these parameters, so it meets the baseline expectation.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly specifies the verb 'Retrieve' and the resource 'detailed screening results', listing contained data (candidate scores, AI reasoning, interview questions). It also adds a constraint ('only completed jobs'), which distinguishes it from sibling tools like get_screening_status.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description explicitly states that the tool is 'only available for completed jobs', providing a clear condition for use. It does not explicitly exclude other tools or list alternatives, but the constraint is actionable and helps the agent decide when to invoke this tool.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

get_screening_statusAInspect

Check the current processing status of a screening job.

ParametersJSON Schema
NameRequiredDescriptionDefault
job_idYesThe unique ID of the job to check.

Output Schema

ParametersJSON Schema
NameRequiredDescription
statusNoCurrent status: processing, completed, or failed.
progressNoPercentage of completion (0-100).
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, and the description only vaguely mentions checking status. It does not confirm that the operation is read-only, safe, or idempotent, nor does it describe possible response statuses.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, concise sentence with no unnecessary words. It effectively communicates the tool's purpose.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the simple parameter and the presence of an output schema (not shown), the description is minimally adequate but lacks additional context such as typical status values or behavior when the job does not exist.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The input schema has 100% coverage with a clear description for the single parameter 'job_id'. The tool description adds no extra meaning beyond what the schema already provides.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the verb 'check' and the resource 'current processing status of a screening job'. It distinguishes itself from siblings like 'get_results' (which returns final results) and 'cancel_screening_job'.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides no explicit guidance on when to use this tool versus alternatives (e.g., 'get_results'). However, the purpose is straightforward, so the omission is not critical.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

list_calendar_connectionsAInspect

List all connected calendar and meeting tools

ParametersJSON Schema
NameRequiredDescriptionDefault

No parameters

Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description must carry the full burden. It correctly identifies the operation as a list (read-only), but lacks details about potential behavioral traits such as pagination, response size, or whether it returns all connections at once.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, concise sentence that is front-loaded with the key action. No unnecessary words, and every part adds value.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a tool with no parameters, no output schema, and no annotations, the description is minimally adequate. It does not clarify the return format or any potential limitations, leaving some ambiguity about what the agent can expect.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The input schema has zero parameters, so schema coverage is effectively 100%. According to guidelines, this earns a baseline of 4. The description adds no parameter info, which is acceptable given no parameters exist.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states 'List all connected calendar and meeting tools', using a specific verb and resource. It effectively distinguishes this tool from siblings like 'create_calendar_connection' and 'generate_meeting_link'.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides no explicit guidance on when to use this tool versus alternatives. While the purpose is clear, it does not mention exclusions or scenarios where other tools would be more appropriate.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

list_credit_packsAInspect

List available credit packs for purchase

ParametersJSON Schema
NameRequiredDescriptionDefault

No parameters

Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

The description indicates a simple read operation without side effects, but lacks details on authentication, freshness, or response structure. No annotations are present to supplement.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, concise sentence with no unnecessary words, efficiently conveying the tool's purpose.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the simplicity of the tool (no parameters, no output schema), the description sufficiently informs the agent of its function. Could optionally mention the return type.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

No parameters exist, so the description does not need to elaborate. Baseline for zero parameters is 4, and the description is adequate.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool lists available credit packs for purchase, which is specific and distinguishes it from siblings like 'purchase_credits' and 'get_credit_balance'.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance is provided on when to use this tool versus alternatives like 'purchase_credits' or 'get_credit_balance'. The context of use is implied but not explicit.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

purchase_creditsCInspect

Purchase credits immediately using saved payment method

ParametersJSON Schema
NameRequiredDescriptionDefault
packNoCredit pack to purchase
amountNoAmount to purchase in dollars (min $5)
payment_method_idYesStripe payment method ID
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations provided, so description must disclose behavioral traits. Only states 'purchase credits immediately' and 'using saved payment method' – lacks details on side effects (e.g., payment charged, balance updated), reversibility, or failure handling.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness3/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Single sentence is concise and front-loaded, but undermines completeness. Every word earns its place, but critical usage details are missing, making it under-specified rather than appropriately concise.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given sibling tools (e.g., create_checkout_session, enable_auto_reload, get_credit_balance), the description fails to clarify its specific role or what it returns. No output schema and no explanation of pack vs amount interaction leaves the agent uncertain.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters2/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100%, but description does not explain relationship between optional pack and amount parameters (e.g., whether both can be specified, or if pack implies a fixed amount). The description adds no value beyond schema for payment_method_id.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

Description clearly states purchase credits immediately using saved payment method. Verb and resource are specific, but does not explicitly differentiate from sibling create_checkout_session, though 'immediately' implies a direct purchase.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance on when to use this tool versus alternatives like create_checkout_session or enable_auto_reload. No prerequisites or exclusions mentioned.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

report_hiring_outcomeCInspect

Report hiring outcome to improve AI accuracy

ParametersJSON Schema
NameRequiredDescriptionDefault
job_idYesThe job ID the candidate belongs to
outcomeYesHiring outcome
candidate_idYesThe candidate ID
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description must disclose behavioral traits. It only says 'report,' which suggests a write operation but does not confirm side effects, return values, or required permissions. The lack of detail leaves the agent uncertain about the tool's behavior.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single sentence with no unnecessary words. It is front-loaded with the verb and resource. While very concise, it sacrifices detail for brevity, which is acceptable for this dimension.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given three required parameters and no output schema or annotations, the description lacks completeness. It does not explain what the tool returns, whether it triggers side effects, or how the outcome is used. The agent lacks sufficient context to predict the tool's full behavior.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The input schema has 100% coverage, so the parameters are documented. The description does not add any additional meaning beyond what the schema provides. For example, it doesn't explain the implications of each outcome enum value. Thus, it meets the baseline but adds no value.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description states the verb 'report' and resource 'hiring outcome' clearly. It also mentions the purpose 'improve AI accuracy', which helps the agent understand the tool's goal. However, it could be more specific about what exactly is reported and to whom.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides no guidance on when to use this tool versus alternatives, nor any conditions for use. There is no mention of prerequisites, exclusions, or preferred contexts.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

test_webhookBInspect

Test a webhook endpoint to verify it is reachable and responding

ParametersJSON Schema
NameRequiredDescriptionDefault
urlYesThe webhook URL to test
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description carries full burden for behavioral disclosure. It only states the tool tests reachability and response, but omits details like whether it sends a test event, expected HTTP response, error handling, timeouts, or side effects.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Single concise sentence with no extraneous information. Every word contributes to the purpose.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a simple one-parameter tool, the description is partially complete. It explains the goal but lacks specifics about what 'responding' entails, potential error states, or expected outcomes. Could be more informative.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100% with a clear description for 'url'. The tool description adds no additional meaning beyond the schema's property description, so baseline 3 is appropriate.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the verb 'Test' and resource 'webhook endpoint', with specific purpose 'verify it is reachable and responding'. This distinguishes it from all sibling tools, none of which test webhooks.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance on when to use this tool vs alternatives, no prerequisites, no conditions to avoid. The description only states what it does, implying usage solely based on the tool name.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

whoamiAInspect

Verify your API token and get profile/balance info

ParametersJSON Schema
NameRequiredDescriptionDefault

No parameters

Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description carries the burden. It discloses that the tool is a read operation ('verify', 'get') and returns profile and balance information, providing adequate behavioral context.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, efficient sentence that conveys all necessary information without waste.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness5/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given no parameters and no output schema, the description fully explains the tool's purpose and what it returns, making it complete for its simplicity.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The tool has zero parameters and schema coverage is 100%, so the description needs no additional parameter information. Baseline 4 applies.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool's function: verify API token and get profile/balance info. It distinguishes itself from sibling tools that focus on email generation, screening, etc.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage for verification and profile/balance retrieval but does not explicitly state when to use this tool versus alternatives or provide exclusions.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

Discussions

No comments yet. Be the first to start the discussion!

Try in Browser

Your Connectors

Sign in to create a connector for this server.

Resources