ukdatapi
Server Details
UK government data intelligence with proprietary scoring from 400+ sources
- Status
- Healthy
- Last Tested
- Transport
- Streamable HTTP
- URL
Glama MCP Gateway
Connect through Glama MCP Gateway for full control over tool access and complete visibility into every call.
Full call logging
Every tool call is logged with complete inputs and outputs, so you can debug issues and audit what your agents are doing.
Tool access control
Enable or disable individual tools per connector, so you decide what your agents can and cannot do.
Managed credentials
Glama handles OAuth flows, token storage, and automatic rotation, so credentials never expire on your clients.
Usage analytics
See which tools your agents call, how often, and when, so you can understand usage patterns and catch anomalies.
Tool Definition Quality
Average 3.8/5 across 25 of 25 tools scored. Lowest: 2.9/5.
Every tool has a clearly distinct purpose, covering different domains like compliance, demographics, property, etc. Descriptions explicitly differentiate overlapping areas, e.g., entity vs. due diligence, location vs. demographics.
Most tools follow the pattern 'uk_{domain}_intelligence', but a few deviate like 'uk_stamp_duty_calculator' and 'uk_vat_validation', causing minor inconsistency.
With 25 tools, the set is borderline heavy. While each tool serves a distinct purpose, the number is at the upper limit of what feels manageable for an agent.
The tool set covers an extensive range of UK data domains with appropriate lifecycle operations, leaving no obvious gaps for a general-purpose UK data API.
Available Tools
25 toolsuk_compliance_intelligenceARead-onlyIdempotentInspect
Assess a UK company's regulatory compliance posture across multiple domains: ICO data protection registration, gender pay gap reporting, modern slavery statements, HSE enforcement notices, environmental permits, and gambling regulation. Returns a Compliance Score (0-100) with EXCELLENT/GOOD/ADEQUATE/CONCERNING/POOR rating and per-domain signals. Use this for pre-acquisition due diligence, supplier compliance checks, or ESG assessments. Companies below regulatory thresholds (e.g., <250 employees for gender pay gap) are scored neutrally, not penalised. For financial risk assessment, use uk_entity_intelligence instead. For director-level risk, use uk_director_intelligence. Sources: ICO, Gender Pay Gap Service, Modern Slavery Registry, HSE, Environment Agency, Gambling Commission.
| Name | Required | Description | Default |
|---|---|---|---|
| depth | No | Detail level. "summary" = ICO + score only (5 credits). "standard" = adds GPG, modern slavery, HSE (15 credits). "full" = adds environmental permits, gambling regulation (30 credits). Default: "standard". | |
| identifier | Yes | Company number (e.g., "00445790") or company name (e.g., "Tesco"). Company numbers give the most precise results. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Annotations already declare the tool read-only and idempotent. The description adds value by detailing the output (score, rating, per-domain signals) and listing data sources, providing transparency beyond the annotations. No contradictions.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is front-loaded with the main purpose and domains, then provides usage guidelines and exclusions. At ~150 words, it is thorough but not overly long. Minor redundancy in listing sources could be trimmed, but overall well-structured.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Despite lacking an output schema, the description fully explains the return format (score and rating) and per-domain signals. It covers purpose, usage, exclusions, threshold handling, and sources. No gaps remain for an agent to correctly select and invoke the tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%, but the description adds meaningful context: it explains that company numbers give more precise results for 'identifier' and specifies credit costs for each 'depth' level ('summary' 5 credits, etc.). This goes beyond the enum values.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool assesses UK company regulatory compliance across multiple specific domains (ICO, gender pay gap, etc.) and returns a Compliance Score with ratings. It distinguishes from siblings 'uk_entity_intelligence' and 'uk_director_intelligence' by specifying alternative tools for financial and director-level risk.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicitly states when to use (pre-acquisition due diligence, supplier checks, ESG) and when not to use (financial risk -> uk_entity_intelligence, director risk -> uk_director_intelligence). Also notes that companies below regulatory thresholds are scored neutrally, avoiding misuse.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
uk_connectivity_intelligenceARead-onlyIdempotentInspect
Evaluate digital infrastructure for any UK postcode. Returns broadband speeds (mean and median), superfast (>30 Mbps) and ultrafast (>100 Mbps) availability, FTTP coverage, mobile 4G indoor and outdoor coverage by operator (EE, Three, O2, Vodafone) with signal strength, and a Digital Readiness Score (0-100). Use this tool to assess remote working viability, evaluate business premises connectivity, or compare digital infrastructure between locations. For physical transport links (rail, bus, roads), use uk_transport_intelligence instead. Source: Ofcom Connected Nations.
| Name | Required | Description | Default |
|---|---|---|---|
| depth | No | Controls response detail. summary: broadband speeds and availability only. standard (default): adds mobile 4G coverage by operator and Digital Readiness Score. full: adds per-operator signal strength breakdown, FTTP availability, and ultrafast metrics. | standard |
| postcode | Yes | Full UK postcode (e.g. "SW1A 1AA"). Connectivity data is returned for the exchange and mobile cell area serving this postcode. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations exist, so the description carries the full burden. It describes the tool as providing intelligence, implying read-only behavior. However, it does not mention authorization needs, rate limits, or any side effects. The description is adequate but lacks depth.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, well-structured sentence that front-loads the tool's purpose and lists key data points. No redundant information. It uses a dash to separate the main function from source attribution.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (2 parameters, no output schema, no nested objects), the description covers the essential inputs and data types. However, it lacks details about the output format (e.g., whether results are structured or text) and any limitations (e.g., postcode format validation).
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%, and the description adds meaning by explaining the 'depth' parameter's levels (summary, standard, full) and their data categories. This goes beyond the schema's brief enum description, helping the agent understand what data each level includes.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states it provides 'UK Digital Connectivity Intelligence' with specific data types (broadband speeds, superfast/ultrafast availability, mobile 4G coverage by operator, Digital Readiness Score). This distinguishes it from sibling tools like uk_demographics_intelligence or uk_location_intelligence, which focus on other domains.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description mentions sources (Ofcom Connected Nations) and lists the data provided, implying use cases for connectivity analysis. However, it does not explicitly state when not to use it or compare it to alternatives like uk_location_intelligence for location-based insights.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
uk_demographics_intelligenceARead-onlyIdempotentInspect
Analyse population demographics and deprivation for any UK postcode at LSOA level. Returns Census 2021 data (population, age, ethnicity, housing tenure, economic activity), Index of Multiple Deprivation with ranks and deciles across all seven domains, Nomis labour market statistics, and a Consumer Spending Power Index (0-100). Use this tool for retail site selection, social impact assessment, or grant applications requiring deprivation evidence. For broader area profiling (crime, flood, food hygiene), use uk_location_intelligence instead. For market opportunity analysis with competitor data, use uk_market_sizing instead. Sources: ONS Census 2021, Nomis, MHCLG IMD.
| Name | Required | Description | Default |
|---|---|---|---|
| depth | No | Controls response detail. summary: labour market statistics only. standard (default): adds Census 2021 population and housing data, IMD ranks, and Consumer Spending Power Index. full: adds detailed Census breakdowns by age band, ethnicity, tenure type, and economic activity category. | standard |
| postcode | Yes | Full UK postcode (e.g. "SW1A 1AA"). Data is returned for the Lower Super Output Area (LSOA) containing this postcode. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the behavioral disclosure burden. It lists key data sources and the summary/standard/full depth levels, giving a sense of scope. However, it does not disclose potential data lag, update frequency, or any restrictions (e.g., valid postcode formats or rate limits).
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, dense paragraph that efficiently conveys the tool's scope, data sources, and a brief note on depth levels. It is front-loaded with the purpose and uses a dash for structure. Only minor excess: the full list of data types could be more structured but is still concise.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no output schema and no annotations, the description provides a reasonable overview of the data included and the depth levels. For a demographics intelligence tool, this level of detail is largely sufficient to inform agent usage, though return structure details could help completeness.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%, and the description does not add much beyond what the schema provides for parameters. The 'depth' parameter is described in the schema with enum values and defaults, and the description only summarizes the depth options. For a tool with good schema coverage, this is acceptable.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool provides UK demographics data from specific sources (ONS Census 2021, Nomis, MHCLG IMD) and includes a proprietary Consumer Spending Power Index. It effectively distinguishes itself from sibling tools which cover different domains (connectivity, education, etc.).
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage for demographic queries but lacks explicit guidelines on when to use this tool versus alternatives. It does not specify prerequisites or limitations, but the context of sibling tools covering different domains provides implicit differentiation.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
uk_director_intelligenceARead-onlyIdempotentInspect
Investigate a company director's full appointment history and risk profile. Returns all current and past directorships with dates, disqualification status, company portfolio analysis (active, dissolved, liquidated counts), Gazette notices, and a Director Risk Score (0-100) based on dissolution rate, disqualification history, and filing compliance. Use this tool to vet a director before appointment, conduct KYC/KYB checks, or investigate connected company networks. For company-level data (not individual directors), use uk_entity_intelligence instead. Sources: Companies House Officers API, Companies House Disqualifications, The Gazette.
| Name | Required | Description | Default |
|---|---|---|---|
| depth | No | Controls response detail. summary: search results, disqualification check, and top 5 appointments. standard (default): adds Gazette notices, full appointment list, and Director Risk Score. full: adds filing compliance check across the director's active company portfolio. | standard |
| name_or_id | Yes | Director name (e.g. "Ken Murphy") for a ranked search, or a Companies House officer ID (e.g. "abc123DEF") for exact lookup. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description carries full burden. It states data sources (Companies House, The Gazette) and lists output components. However, it doesn't disclose if this is a read-only operation, if it returns cached vs live data, or any rate limits. With no annotations, this is a moderate disclosure but insufficient for full transparency.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Description is concise (3 sentences) with front-loaded purpose and structured detail. Every sentence adds value. Could be slightly improved by a briefer second sentence, but overall efficient.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given 2 params, no output schema, and no annotations, the description provides adequate coverage: input semantics (name_or_id), depth variants, and output components. It explains what the tool returns but does not specify return format or response structure. However, for a search/info tool without output schema, this is reasonably complete.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100% with detailed descriptions for both parameters. Description adds context about what each depth level returns ('summary', 'standard', 'full'), which significantly complements the enum options in schema. However, description doesn't elaborate beyond what's already in schema param descriptions.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description clearly states it provides UK Director Network Intelligence, specifying search by name or officer ID. It lists all returned data types including disqualifications, portfolio analysis, Gazette notices, and risk score. This is highly specific and distinguishes it from sibling tools which focus on different UK intelligence domains.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Description implies when to use (searching directors), but provides no explicit guidance on when not to use or how it differs from sibling tools like uk_entity_intelligence. The depth parameter gives usage variants, but no advice on selecting between them.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
uk_due_diligence_reportARead-onlyIdempotentInspect
Generate a company due diligence report with an opinionated verdict: PROCEED, PROCEED_WITH_CAUTION, ENHANCED_DUE_DILIGENCE, or DO_NOT_ENGAGE. Returns a Corporate Distress Score (0-100), categorised red and green flags with severity ratings, governance quality assessment, beneficial ownership transparency check, and an actionable recommendation. Use this tool for a quick go/no-go decision on a supplier, partner, or acquisition target. For raw company data without a verdict, use uk_entity_intelligence instead. Sources: Companies House, FCA Register, Charity Commission, The Gazette, Contracts Finder.
| Name | Required | Description | Default |
|---|---|---|---|
| identifier | Yes | Company number (e.g. "00000006") or company name. Numbers are matched exactly; names trigger a ranked search. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It discloses key behavioral traits: the tool generates an opinionated verdict, includes specific metrics (Corporate Distress Score, red/green flags), and sources data from listed authorities. However, it lacks details on rate limits, authentication needs, or error handling, leaving gaps in operational context.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is front-loaded with the core purpose and structured efficiently in two sentences, covering verdicts, components, cost replacement, and sources. It avoids unnecessary details, though it could be slightly more concise by integrating some elements more tightly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (synthesizing reports from multiple sources) and lack of annotations or output schema, the description is moderately complete. It outlines the report's components and sources but does not detail the output structure, potential limitations, or integration specifics, leaving room for improvement in guiding the agent.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, with the parameter 'identifier' documented as 'Company number or company name'. The description adds no additional parameter semantics beyond what the schema provides, such as format examples or validation rules, so it meets the baseline for high schema coverage without extra value.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: synthesizing a company due diligence report with a specific verdict system (PROCEED, PROCEED_WITH_CAUTION, ENHANCED_DUE_DILIGENCE, DO_NOT_ENGAGE). It distinguishes itself from sibling tools by focusing on due diligence rather than education, energy, legal, or other intelligence domains, and specifies it replaces specific CreditSafe reports.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage context by stating it synthesizes due diligence reports for UK companies, but does not explicitly state when to use this tool versus alternatives or provide exclusions. It mentions sources (Companies House, FCA, etc.), which suggests applicability to UK entities, but lacks direct guidance on tool selection among siblings.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
uk_education_intelligenceARead-onlyIdempotentInspect
Find schools near any UK postcode with Ofsted ratings, school type (academy, maintained, free, independent), phase, age range, pupil numbers, and distance. Results are sorted by proximity. Use the phase parameter to filter to primary or secondary schools only. Use this tool for relocation advice, catchment area research, or comparing local education quality. This tool covers schools only; for broader area profiling, use uk_location_intelligence. Sources: DfE Get Information About Schools (GIAS), Ofsted Inspection Outcomes.
| Name | Required | Description | Default |
|---|---|---|---|
| depth | No | Controls response detail and search radius. summary: nearest 5 schools within 1 km. standard (default): schools within 2 km with Ofsted ratings. full: extended catchment area with detailed school profiles. | standard |
| phase | No | Optional filter by school phase. When omitted, all phases are returned. | |
| postcode | Yes | Full UK postcode (e.g. "SW1A 1AA"). Schools are returned sorted by distance from this postcode. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It mentions the data source (DfE GIAS + Ofsted), which adds context about reliability and scope, but lacks details on rate limits, authentication needs, error handling, or response format. The description doesn't contradict annotations, but could be more informative about behavioral traits.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is front-loaded and concise, using a single sentence that efficiently conveys the tool's purpose, scope, and data sources without unnecessary details. Every word earns its place, making it easy to understand quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (3 parameters, no output schema, no annotations), the description is adequate but has gaps. It covers the purpose and data sources but lacks information on output format, error cases, or usage constraints. Without annotations or output schema, more behavioral context would improve completeness.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents parameters well. The description adds value by clarifying the tool's focus on schools and data sources, but doesn't provide additional parameter semantics beyond what's in the schema. With 3 parameters and high schema coverage, a baseline of 3 is appropriate, but the description's context slightly enhances understanding.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with specific verbs ('schools near any postcode') and resources (Ofsted ratings, type, phase, pupil numbers), and distinguishes itself from siblings by focusing on education intelligence rather than other domains like property, transport, or health.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage by mentioning 'schools near any postcode,' suggesting it's for location-based queries, but provides no explicit guidance on when to use this tool versus alternatives or any exclusions. It doesn't reference sibling tools or specify scenarios where other tools might be more appropriate.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
uk_energy_intelligenceBRead-onlyIdempotentInspect
Analyse energy infrastructure and sustainability for any UK postcode. Returns real-time grid carbon intensity (gCO2/kWh) with forecast, electricity generation mix (wind, solar, gas, nuclear percentages), wholesale prices, local EPC rating distribution, and an expanded ESG Assessment Score (0-100, rated STRONG/GOOD/MODERATE/WEAK/POOR) with weighted component breakdown covering carbon intensity, renewable share, and building efficiency. Use this tool for ESG reporting, renewable energy viability, or comparing sustainability between locations. For environmental risk factors (flood, geology, radon), use uk_environmental_risk instead. Sources: National Grid ESO, Elexon BMRS, EPC Register.
| Name | Required | Description | Default |
|---|---|---|---|
| depth | No | Controls response detail. summary: current carbon intensity only. standard (default): adds generation mix, local EPC profile, and ESG score. full: adds wholesale energy prices, carbon intensity forecast, and detailed EPC breakdown. | standard |
| postcode | Yes | Full UK postcode (e.g. "SW1A 1AA"). Energy data is returned for the grid region and local area. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. While it lists the types of data returned and sources, it doesn't describe important behavioral aspects like whether this is a read-only operation, rate limits, authentication requirements, data freshness, or error conditions. The description provides content scope but lacks operational context.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely efficient - a single sentence packed with essential information: tool scope, specific data types, and data sources. Every element earns its place with zero wasted words. The structure is front-loaded with the core purpose followed by implementation details.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a tool with 2 parameters, 100% schema coverage, but no annotations or output schema, the description provides adequate context about what data is returned and sources. However, it lacks information about return format, error handling, and operational constraints that would be important for an AI agent to use this tool effectively. The completeness is minimal but viable.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
With 100% schema description coverage, the baseline is 3. The description adds value by implicitly explaining what data corresponds to each depth level: carbon intensity (summary), generation mix and EPC (standard), wholesale prices and ESG score (full). This provides semantic context beyond the schema's technical descriptions, though it doesn't fully explain the 'postcode' parameter's role beyond what the schema states.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool provides UK energy and ESG intelligence including specific data types like grid carbon intensity, generation mix, wholesale prices, local EPC efficiency, and ESG scores. It distinguishes itself from siblings by focusing specifically on energy/ESG data rather than other domains like education, health, or transport. However, it doesn't explicitly state the action verb (e.g., 'retrieve' or 'get') that the tool performs.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus the sibling tools. While it's clear this tool focuses on energy/ESG data, there's no explicit comparison to alternatives like 'uk_environmental_risk' or 'uk_property_intelligence' that might overlap in some domains. No prerequisites, limitations, or specific use cases are mentioned.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
uk_entity_intelligenceARead-onlyIdempotentInspect
Look up any UK company, charity, or FCA-regulated entity by company number, charity number, or name. Returns company profile, officers, persons of significant control, filings, charges, FCA and Charity Commission regulatory status, Gazette notices, government contracts, and a Corporate Distress Score (0-100). Use this tool to verify a company, check its status, identify directors, or assess financial stability. For a go/no-go verdict on a company, use uk_due_diligence_report instead. To investigate a specific director's history across companies, use uk_director_intelligence instead. Sources: Companies House, FCA Register, Charity Commission, The Gazette, Contracts Finder.
| Name | Required | Description | Default |
|---|---|---|---|
| depth | No | Controls response detail. summary: basic profile only. standard (default): adds officers, regulatory status, distress score. full: adds government contracts, officer network analysis, and complete filing history. | standard |
| identifier | Yes | Company number (e.g. "00000006"), charity number (e.g. "1234567"), or company name. Numbers are matched exactly; names return the best ranked match. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden. It discloses data sources (Companies House, FCA, etc.) and output types (profile, officers, risk signals), but lacks details on rate limits, authentication needs, error handling, or pagination. It adequately describes what the tool returns but misses operational constraints.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is front-loaded with the core purpose and efficiently lists returned data and sources in two sentences. Every sentence adds value, though it could be slightly more structured (e.g., separating data points from sources). No wasted words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no annotations and no output schema, the description provides good coverage of what the tool does and returns. However, for a tool with rich data outputs and multiple parameters, it lacks details on response format, error cases, or limitations, leaving gaps in completeness for an AI agent.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the baseline is 3. The description adds value by clarifying that 'identifier' accepts company numbers, charity numbers, or company names, which enriches the schema's generic description. However, it does not explain the semantic impact of 'depth' levels beyond what the schema already details.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the action ('Look up'), the target resources ('UK company, charity, or regulated entity'), and provides a comprehensive list of returned data points. It distinguishes itself from siblings by focusing on entity intelligence rather than due diligence, education, energy, etc., which are domain-specific.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage for entity lookup but does not explicitly state when to use this tool versus alternatives like 'uk_due_diligence_report' or other sibling tools. No guidance on exclusions or prerequisites is provided, leaving usage context inferred rather than defined.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
uk_environmental_riskARead-onlyIdempotentInspect
Calculate a multi-factor Environmental Risk Score (0-100) for any UK postcode. Combines flood zone classifications (rivers, surface, coastal), ground stability and shrink-swell clay hazard from BGS, radon gas probability, waterbody ecological status under the Water Framework Directive, and grid carbon intensity. Returns individual risk component scores with severity ratings and an overall weighted composite. Use this tool for land development evaluation, insurance exposure assessment, or environmental due diligence. For water-specific pollution and sewage data, use uk_water_intelligence instead. Sources: Environment Agency, BGS GeoIndex, EA Water Quality Archive, National Grid ESO.
| Name | Required | Description | Default |
|---|---|---|---|
| depth | No | Controls response detail. summary: flood risk and geology only. standard (default): adds water quality classification and composite score. full: adds carbon intensity, detailed flood monitoring station readings, and full risk component breakdown. | standard |
| postcode | Yes | Full UK postcode (e.g. "SW1A 1AA"). Returns environmental risk data for the surrounding area. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It effectively describes what the tool calculates and its data sources, but doesn't disclose important behavioral traits like rate limits, authentication requirements, response format, error handling, or whether it's a read-only operation. The description doesn't contradict any annotations since none exist.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is perfectly concise and well-structured - a single sentence that efficiently communicates the core functionality, scoring range, geographic scope, specific factors, and data sources. Every element earns its place with no wasted words, and the information is front-loaded with the most important details first.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (2 parameters, no output schema, no annotations), the description provides a solid foundation but has gaps. It explains what the tool does and its data sources well, but doesn't describe the output format, error conditions, or operational constraints. Without annotations or output schema, the agent lacks complete information about what to expect from this tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
With 100% schema description coverage, the baseline is 3. The description adds value by explaining what factors are included at each depth level ('summary = flood + geology. standard = adds water quality. full = adds carbon intensity'), which provides important context beyond the schema's enum values. However, it doesn't elaborate on the postcode parameter beyond what the schema already states.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: calculating a multi-factor environmental risk score (0-100) for UK postcodes, specifying the exact factors included (flood risk, ground stability, shrink-swell clay, radon potential, water quality, carbon intensity) and data sources. It distinguishes itself from sibling tools by focusing specifically on environmental risk assessment rather than education, health, legal, or other intelligence domains.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage context through the listed data sources and factors, suggesting it's for environmental risk assessment in the UK. However, it doesn't explicitly state when to use this tool versus alternatives or provide any exclusion criteria. The depth parameter description offers some implicit guidance on output detail levels but no explicit usage rules.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
uk_funding_intelligenceARead-onlyIdempotentInspect
Discover UK grants, research funding, and government contracts matching a search query. Returns grants from the 360Giving registry (funder, amount, recipient), UKRI-funded research projects (PI, institution), government contracts from Contracts Finder, and a Funding Opportunity Score (0-100). Use this tool to find grant funding, identify research collaboration opportunities, or map the funding landscape for a sector. For live government procurement tenders, use uk_tenders_intelligence instead. For company patent and R&D data, use uk_innovation_intelligence instead. Sources: 360Giving, UKRI Gateway to Research, Contracts Finder.
| Name | Required | Description | Default |
|---|---|---|---|
| depth | No | Controls response detail. summary: matching grants only. standard (default): adds UKRI research projects, government contracts, and Funding Opportunity Score. full: adds detailed grant breakdowns and funder profiles. | standard |
| query | Yes | Sector, topic, or keyword (e.g. "clean energy", "AI healthcare", "social housing innovation"). Broader terms return more results across all funding sources. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so the description must convey behavior. It mentions the data sources and scoring, but does not disclose if the tool creates or modifies data, if there are rate limits, or what the output format looks like. The description is adequate for a search tool but lacks depth.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is concise (two sentences) with key information front-loaded, but could be slightly more structured (e.g., bullet points for depth levels). No wordiness.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the simple two-parameter schema and no output schema, the description adequately covers what the tool does and its inputs. However, absence of output format hints reduces completeness slightly.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100% with descriptions for both parameters. The description adds context by explaining the depth levels and giving example queries, providing additional meaning beyond the schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: searching UK grants, funding, and related opportunities from multiple sources (360Giving, UKRI, Contracts Finder). It mentions key features like a Funding Opportunity Score and search dimensions (sector, topic, keyword), and distinguishes it from sibling tools that focus on other UK-specific domains.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies use cases via examples ('clean energy', 'AI') and explains the depth parameter options, but does not explicitly state when to use this tool versus alternatives (e.g., uk_tenders_intelligence for contracts only). There is no 'when-not-to-use' guidance.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
uk_health_intelligenceCRead-onlyIdempotentInspect
Map the healthcare landscape around any UK postcode. Returns public health indicators (life expectancy, obesity, smoking rates, mental health, deprivation health rank), CQC-inspected providers with ratings and inspection dates (hospitals, GPs, care homes, dentists, pharmacies), GP prescribing patterns, and a Health Service Quality Score (0-100, rated EXCELLENT/GOOD/MODERATE/POOR/INADEQUATE). Use this tool for health impact assessments, care home due diligence, or evaluating local health service quality. For environmental health risks (flood, radon, pollution), use uk_environmental_risk instead. Sources: OHID Fingertips, CQC, OpenPrescribing.
| Name | Required | Description | Default |
|---|---|---|---|
| depth | No | Controls response detail. summary: key public health indicators only. standard (default): adds CQC-rated providers with inspection outcomes. full: adds GP prescribing data and detailed health indicator breakdowns. | standard |
| postcode | Yes | Full UK postcode (e.g. "SW1A 1AA"). Healthcare data is returned for the surrounding area. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It describes what data is available (health indicators, CQC providers, prescribing patterns) but doesn't disclose important behavioral traits like whether this is a read-only operation, potential rate limits, authentication requirements, data freshness, or what format/scope the results will have. For a data query tool with no annotation coverage, this leaves significant gaps.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise and well-structured in a single sentence that packs substantial information: the tool's domain (UK healthcare intelligence), three specific data types it provides, the geographic scope (any postcode), and three data sources. Every element earns its place with zero wasted words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a data query tool with no annotations and no output schema, the description is incomplete. While it efficiently states what data is available, it doesn't address how results are structured, whether there are limitations (like postcode validation requirements), data freshness, or error conditions. Given the complexity of healthcare intelligence data and lack of structured metadata, more contextual information would be helpful.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents both parameters thoroughly. The description doesn't add any parameter semantics beyond what's in the schema - it mentions postcodes and different data types but doesn't explain parameter interactions or provide additional context about how parameters affect results. Baseline 3 is appropriate when the schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool provides UK healthcare intelligence including health indicators, CQC-rated care providers, and GP prescribing patterns for any postcode, with specific data sources listed. It uses specific verbs ('provides') and resources (health indicators, CQC providers, prescribing patterns), though it doesn't explicitly differentiate from sibling tools beyond the healthcare domain focus.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. While it's clear this is for UK healthcare intelligence, there's no mention of when to choose this over other intelligence tools (like education_intelligence or environmental_risk) or what specific healthcare questions it addresses best. No exclusions or prerequisites are stated.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
uk_innovation_intelligenceARead-onlyIdempotentInspect
Assess a company's innovation activity and IP portfolio by name or Companies House number. Returns EPO patent filings and grants with IPC codes, UKRI research grants with funding amounts, R&D intensity analysis by SIC code, and an Innovation Score (0-100). Use this tool for investment due diligence on IP strength, benchmarking innovation against peers, or identifying patent-active competitors. For company financial health and governance data, use uk_entity_intelligence instead. For grant funding opportunities (not company-specific), use uk_funding_intelligence instead. Sources: EPO Open Patent Services, UKRI Gateway to Research, Companies House.
| Name | Required | Description | Default |
|---|---|---|---|
| depth | No | Controls response detail. summary: patent count and SIC-based R&D analysis only. standard (default): adds UKRI research grants, full patent list, and Innovation Score. full: adds detailed patent abstracts, IPC breakdowns, and grant funding amounts. | standard |
| identifier | Yes | Company number (e.g. "00000006") or company name. Numbers give exact results; names trigger a search. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Annotations are absent, so the description carries the full burden. It discloses data sources and the Innovation Score range, but does not mention behavioral aspects such as data freshness, rate limits, or potential delays. This leaves some uncertainty about the tool's performance.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is 3 sentences long and front-loaded with the tool's purpose. It is concise, but the depth level details could be more integrated rather than appended as a separate sentence.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (2 params, no output schema), the description covers the key aspects: purpose, inputs, depth variations, and data sources. It does not describe return values, but without an output schema, the agent might need to infer what is returned.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
With 100% schema description coverage, the baseline is 3. The description adds value by explaining the depth levels: 'summary = patents + SIC analysis', 'standard = adds UKRI research + full scoring', 'full = adds detailed patent analysis'. This clarifies the semantic differences beyond the enum labels.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool provides UK innovation intelligence with specific data sources (EPO, UKRI, Companies House) and components (patents, grants, SIC analysis, Innovation Score). It also specifies the input required (company name or number), making it easy to understand the tool's purpose and distinguish it from siblings.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage: search by company name or number. However, it does not provide explicit guidance on when to use this tool versus alternatives like other intelligence tools, nor does it mention when not to use it.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
uk_legal_intelligenceBRead-onlyIdempotentInspect
Search UK legislation and track pending Parliamentary bills by topic, sector, or act name. Returns matching acts of Parliament and statutory instruments with title, year, type, and legislation.gov.uk link, plus active bills with their current parliamentary stage and sponsoring department, and a Regulatory Relevance Score (0-100, rated HIGHLY_REGULATED/REGULATED/MODERATE/LIGHTLY_REGULATED/MINIMAL). Returns up to 10 legislation items at standard depth and 20 at full depth. Use this tool for regulatory research, compliance checks, or tracking bills. For government procurement contracts, use uk_tenders_intelligence instead. For grants and research funding, use uk_funding_intelligence instead. Sources: legislation.gov.uk, UK Parliament Bills API.
| Name | Required | Description | Default |
|---|---|---|---|
| depth | No | Controls response detail. summary: top 3 matching legislation items only, no bills. standard (default): up to 10 legislation items plus relevant pending bills. full: up to 20 legislation items plus all currently active bills. | standard |
| query | Yes | Topic keyword (e.g. "data protection"), sector name (e.g. "financial services"), or specific act (e.g. "Companies Act 2006"). Broader terms return more results. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It mentions sources but doesn't disclose behavioral traits like whether this is a read-only search, rate limits, authentication needs, or what the output format looks like. For a search tool with zero annotation coverage, this leaves significant gaps in understanding how it behaves.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is appropriately sized with two sentences that efficiently convey purpose and sources. It's front-loaded with the main function and avoids unnecessary details, though it could be slightly more structured by separating usage hints.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (search with one parameter), no annotations, and no output schema, the description is minimally adequate. It covers what the tool does and sources but lacks details on behavior, output, or error handling, leaving room for improvement in completeness.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The description doesn't add parameter-specific information beyond what the input schema provides. With 100% schema description coverage for the single 'query' parameter, the baseline is 3. The description generalizes about searching sectors/topics but doesn't elaborate on query syntax, examples beyond the schema, or constraints.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool searches UK legislation, pending Parliamentary bills, and relevant acts for any sector or topic, specifying the verb 'search' and resources. It distinguishes from sibling tools by focusing on legal/policy intelligence rather than education, energy, property, etc., though it doesn't explicitly contrast with them.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage for legal/policy topics but doesn't explicitly state when to use this tool versus alternatives like uk_entity_intelligence or uk_due_diligence_report. It mentions sources (legislation.gov.uk, Parliament Bills API), which provides some context but no clear exclusions or comparative guidance.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
uk_location_intelligenceARead-onlyIdempotentInspect
Profile any UK postcode with safety, environmental, and socioeconomic data. Returns admin hierarchy, deprivation rank, 12-month crime statistics by category, flood risk zones, food hygiene ratings, carbon intensity, labour market indicators, and health metrics. Use this tool for area assessment, relocation decisions, or neighbourhood comparison. For detailed Census demographics and IMD breakdowns, use uk_demographics_intelligence instead. For property-specific data (prices, EPC), use uk_property_intelligence instead. Sources: ONS, Police.uk, Environment Agency, Food Standards Agency, Nomis, OHID Fingertips.
| Name | Required | Description | Default |
|---|---|---|---|
| depth | No | Controls response detail. summary: postcode lookup and admin hierarchy only. standard (default): adds crime statistics, flood risk, food hygiene. full: adds carbon intensity, labour market data, and health indicators. | standard |
| postcode | Yes | Full UK postcode (e.g. "SW1A 1AA" or "SW1A1AA"). Partial postcodes are not supported. | |
| radius_m | No | Search radius in metres for nearby data such as crime and food hygiene. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It mentions data sources (ONS, Police.uk, EA, FSA, Nomis, Fingertips) which adds some context about reliability and scope, but it doesn't disclose behavioral traits like rate limits, authentication needs, response formats, error conditions, or whether the operation is read-only or has side effects. For a tool with no annotation coverage, this leaves significant gaps.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is efficiently structured in two sentences: the first states the purpose and comprehensive return data, the second lists data sources. Every element adds value without redundancy, and it's front-loaded with the core functionality. No wasted words or unnecessary details.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity (profiling tool with multiple data types), no annotations, and no output schema, the description is moderately complete. It covers what data is returned and sources, but lacks details on output format, error handling, or behavioral constraints. For a tool with rich data integration but no structured output documentation, this leaves room for improvement in guiding the agent effectively.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all parameters well. The description doesn't add any parameter-specific information beyond what's in the schema, but since there are only 3 parameters and the schema is comprehensive, this is acceptable. The baseline would be 3, but the description's listing of returned data types implicitly reinforces the 'depth' parameter's purpose, earning a slightly higher score.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the specific action ('Profile any UK postcode') and lists the comprehensive data returned (admin hierarchy, demographics, deprivation indices, crime statistics, flood risk, food hygiene, carbon intensity, labour market, and health indicators). It distinguishes itself from siblings by focusing on postcode-based profiling rather than entity intelligence, property intelligence, or other domain-specific tools.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage context by specifying 'UK postcode' and listing data sources, but it doesn't explicitly state when to use this tool versus alternatives like uk_property_intelligence or uk_environmental_risk. No guidance is provided about prerequisites, limitations, or specific scenarios where this tool is preferred over siblings.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
uk_market_sizingARead-onlyIdempotentInspect
Analyse market opportunity for any UK postcode with optional sector filtering. Returns a Market Opportunity Score (0-100), labour market data (earnings, employment rates, claimant count), property market context (average prices, transaction volumes, price trends), and competitor density by SIC code. When a sector keyword is provided, returns counts of local competing businesses. Use this tool for feasibility studies, franchise territory analysis, or comparing commercial potential between locations. For Census demographics and deprivation data, use uk_demographics_intelligence instead. Sources: ONS Census 2021, Nomis, HM Land Registry, Companies House.
| Name | Required | Description | Default |
|---|---|---|---|
| depth | No | Controls response detail. summary: labour market data only. standard (default): adds property market context, competitor counts, and opportunity score. full: adds detailed competitor company profiles and extended market indicators. | standard |
| sector | No | Optional SIC code or sector keyword for competitor analysis (e.g. "62020", "restaurant", "construction"). When omitted, competitor analysis is skipped. | |
| postcode | Yes | Full UK postcode (e.g. "M1 1AA"). Market analysis is centred on this location. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It mentions data sources (ONS Census, Nomis, etc.) and output types, but fails to disclose critical behavioral traits such as rate limits, authentication needs, data freshness, error handling, or whether the analysis is cached or real-time. For a data-intensive tool with no annotation coverage, this is a significant gap.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is efficiently structured in two sentences: the first states the purpose and outputs, the second lists data sources. Every sentence adds essential information with zero waste, making it easy to parse and front-loaded with key details.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (market analysis with multiple data sources) and no output schema, the description is moderately complete. It outlines what the tool returns but lacks details on output format, structure, or example responses. With no annotations and rich functionality, it should provide more behavioral context to be fully adequate.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema documents all parameters well. The description adds value by explaining the optional 'sector' parameter's purpose ('for competitor search') and providing examples ('62020' for IT consultancy, 'restaurant'), which enhances understanding beyond the schema's basic descriptions.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with specific verbs ('analysis', 'returns') and resources ('UK postcode', 'sector'), detailing the exact outputs (Market Opportunity Score, labour market data, property market context, competitor density). It distinguishes itself from siblings by focusing on market sizing rather than due diligence, education, energy, or other intelligence types.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage for market opportunity analysis in the UK, but provides no explicit guidance on when to use this tool versus alternatives like 'uk_location_intelligence' or 'uk_property_intelligence'. It mentions optional sector filtering but lacks context on prerequisites or exclusions.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
uk_planning_intelligenceARead-onlyIdempotentInspect
Planning and development intelligence for any UK postcode. Returns planning application history (approved/refused/pending/withdrawn counts), conservation areas, listed buildings by grade, Green Belt status, Article 4 directions, brownfield data, flood risk, council approval rates with portal URL, environmental designations (SSSI, AONB), mining risk, and a Development Score (0-100) from HIGHLY_CONSTRAINED to HIGHLY_FAVOURABLE. Use this tool to assess planning permission potential or prepare application strategy. For property-level data (prices, EPC), use uk_property_intelligence instead. Sources: Planning Data Platform, Natural England, MHCLG, Environment Agency, Coal Authority.
| Name | Required | Description | Default |
|---|---|---|---|
| depth | No | Controls response detail. summary: key constraints, conservation area, flood risk, and development score. standard (default): adds planning application breakdown, environmental designations, council stats with portal URL. full: adds mining risk, detailed brownfield data, and full application history. | standard |
| postcode | Yes | Full UK postcode (e.g. "SW1A 1AA"). Returns planning data for the surrounding area. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description carries full burden. It discloses data sources and the output score range (0-100) and categories. However, it doesn't state whether the tool is read-only or any rate limits, which is a minor gap.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Description is efficient at 3 sentences, but the first sentence is a bit long listing many items. Still, it front-loads the purpose and provides comprehensive details without fluff.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no output schema, description adequately covers return values by listing the main categories. Depth parameter explanation compensates for the missing output schema. However, it could mention that results are for a single postcode and any pagination limits.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so baseline is 3. Description adds value by detailing what each depth level returns (summary=constraints+score, standard=adds applications, designations, council stats, full=adds mining risk, brownfield, full history), going beyond the enum labels.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description clearly states the tool returns UK planning intelligence for any postcode, listing specific data types (planning applications, conservation areas, etc.) and the proprietary Planning Development Score. It distinguishes from siblings like uk_property_intelligence and uk_location_intelligence by focusing on planning-specific data.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Description implies use for planning-related queries by listing relevant data sources (Planning Data Platform, Natural England). No explicit when-not-to-use or alternatives among siblings, but the scope is clear from the data types.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
uk_political_intelligenceARead-onlyIdempotentInspect
Get political representation and civic engagement data for any UK postcode. Returns the current MP with party affiliation, majority size, and tenure, active parliamentary petitions with constituency-level signature counts, and a Political Engagement Index measuring civic participation relative to the national average. Use this tool for political landscape research, public sentiment gauging via petition activity, or stakeholder engagement preparation. Sources: UK Parliament Members API, UK Parliament Petitions API.
| Name | Required | Description | Default |
|---|---|---|---|
| depth | No | Controls response detail. summary: current MP profile only. standard (default): adds top petitions with local signature counts and Political Engagement Index. full: adds detailed petition analysis and historical MP data. | standard |
| postcode | Yes | Full UK postcode (e.g. "SW1A 1AA"). Returns data for the parliamentary constituency containing this postcode. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It cites data sources (UK Parliament Members API, UK Parliament Petitions) which adds transparency, but does not disclose important behavioral traits such as authorization requirements, rate limiting, or whether the tool is read-only.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is three sentences, each with a clear purpose: first sentence states what the tool retrieves, second mentions data sources, and overall structure is efficient with no wasted words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a tool with no output schema, the description should ideally mention the return format or structure. It names the key outputs but gives no indication of whether results are paginated, or how to interpret the 'Political Engagement Index'. Coverage is adequate but not complete.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100% and the description clarifies the 'depth' enum beyond the schema, explaining what each level returns (summary/standard/full). The postcode parameter is briefly described. This adds meaningful context beyond the schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description specifies exactly what the tool retrieves: current MP with party and tenure, active parliamentary petitions with constituency-level signature counts, and Political Engagement Index. It clearly distinguishes from siblings by being UK-specific and focused on political data.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description indicates the tool is for UK political intelligence, but does not explicitly state when to use it vs alternatives. Siblings cover other domains (connectivity, demographics, etc.), but no direct guidance is given.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
uk_property_intelligenceARead-onlyIdempotentInspect
Property due diligence for any UK postcode. Returns HM Land Registry price-paid history with area averages, EPC energy performance certificates (90+ fields including ratings, wall/roof/heating descriptions, CO2 emissions), planning constraints, and an Environmental Risk Score (0-100) combining flood, ground stability, radon, and crime. Use this tool for property purchase research, conveyancing preparation, buy-to-let analysis, or rental market assessment. For broader area profiling (demographics, schools, transport), use uk_location_intelligence instead. Sources: HM Land Registry, EPC Register, Planning Data Platform, Environment Agency, BGS, Police.uk.
| Name | Required | Description | Default |
|---|---|---|---|
| depth | No | Controls response detail. summary: price history and flood risk only. standard (default): adds EPC data, planning constraints, crime statistics. full: adds geology, radon, and detailed environmental risk scoring with component breakdown. | standard |
| postcode | Yes | Full UK postcode (e.g. "SW1A 1AA"). Returns data for all properties in this postcode area. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden. It discloses behavioral traits such as the comprehensive data returned (price history, EPC, planning, environmental risk) and sources (Land Registry, etc.), but it lacks details on rate limits, authentication needs, or error handling, which are important for a tool with such broad data aggregation.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is front-loaded with the core purpose, followed by specific return data and context, all in a single, efficient sentence with no wasted words. Every part adds value, making it highly concise and well-structured.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (aggregating multiple data sources) and no output schema, the description provides a good overview of return values but lacks specifics on format, pagination, or error cases. With no annotations, it should do more to cover behavioral aspects like data freshness or limitations.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 100% description coverage, so the baseline is 3. The description adds value by explaining the 'depth' parameter's impact on output (e.g., 'summary = prices + flood'), which clarifies semantics beyond the schema's enum descriptions, though it doesn't detail the 'postcode' parameter further.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with specific verbs ('Complete property due diligence') and resources ('UK postcode'), and it distinguishes from siblings by focusing on comprehensive property intelligence rather than specialized areas like education, energy, or transport covered by other tools.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage context by mentioning it 'Replaces £20-50 conveyancing searches,' suggesting it's for property assessment, but it doesn't explicitly state when to use this tool versus alternatives like 'uk_environmental_risk' or 'uk_location_intelligence,' which might overlap in some data areas.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
uk_stamp_duty_calculatorARead-onlyIdempotentInspect
Calculate UK Stamp Duty Land Tax (SDLT) for England and Northern Ireland property purchases. Returns total tax due, effective tax rate, and a breakdown by band. Includes first-time buyer relief (nil rate up to £425,000 on properties up to £625,000) and additional property surcharge (+5% for buy-to-let and second homes) at 2025/26 rates. Use this tool when a user asks about stamp duty, SDLT, property purchase tax, or how much tax they'll pay on a house purchase. For full property due diligence including EPC, flood risk, and planning constraints, use uk_property_intelligence instead. Source: HMRC SDLT Rates.
| Name | Required | Description | Default |
|---|---|---|---|
| price | Yes | Property purchase price in GBP (e.g., 350000 for a £350,000 property). | |
| buyer_type | No | Buyer category. standard: normal purchase. first-time: first-time buyer (eligible for relief on properties up to £625,000). additional: additional property such as buy-to-let or second home (+5% surcharge). | standard |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Annotations already declare readOnlyHint=true, idempotentHint=true, destructiveHint=false. Description adds important behavioral details: includes first-time buyer relief, additional property surcharge, 2025/26 rates, and HMRC source. No contradictions with annotations.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Four well-structured sentences with no waste. Front-loaded with purpose, followed by output details, usage guide, and alternative tool mention.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
No output schema, but description sufficiently explains return values (total tax, effective rate, breakdown). Covers tax year and source, providing complete context for a calculator tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%, but description adds specific relief details (nil rate up to £425k for first-time buyers on properties up to £625k) and surcharge (+5%) that are not in schema parameter descriptions, enhancing meaning.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description clearly states the tool calculates UK Stamp Duty Land Tax for England and Northern Ireland, specifying what it returns (total tax, effective rate, breakdown). It distinguishes from sibling uk_property_intelligence by mentioning it covers full property due diligence.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicitly states when to use (when user asks about stamp duty, SDLT, property purchase tax) and when not to use (for full due diligence, use uk_property_intelligence), providing clear guidance and an alternative.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
uk_tenders_intelligenceCRead-onlyIdempotentInspect
Search open UK government procurement opportunities and recently awarded public contracts from the last 90 days. Returns up to 30 tenders with title, contracting authority, estimated or awarded value in GBP, deadline, status (open, awarded, closed), and a Procurement Opportunity Score (0-100, rated EXCELLENT/STRONG/MODERATE/LIMITED/POOR) assessing opportunity quality. Both parameters are optional; when omitted, returns the most recent tenders across all sectors and regions. Use this tool to find public sector sales opportunities, track government spending, or research competitor contract wins. For grants and research funding (non-procurement), use uk_funding_intelligence instead. For import/export duties, use uk_trade_intelligence instead. Sources: Contracts Finder, Find a Tender Service (FTS).
| Name | Required | Description | Default |
|---|---|---|---|
| depth | No | Controls response detail. summary: top 5 tenders with summary statistics only. standard (default): all matching tenders with buyer and deadline details. full: adds direct links to tender notices and value distribution analysis. | standard |
| region | No | UK region to filter tenders (e.g. "London", "North West", "Scotland", "Wales"). When omitted, returns tenders from all regions. | |
| sector | No | Sector keyword to filter tenders (e.g. "IT", "construction", "healthcare", "facilities management"). When omitted, returns tenders across all sectors. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden for behavioral disclosure. It states the tool provides 'intelligence' but doesn't clarify whether this is a read-only search, requires authentication, has rate limits, returns structured data, or involves any destructive operations. The description is too vague about actual behavior, leaving significant gaps for agent understanding.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is efficiently structured in two sentences: first stating the purpose and data types, second listing sources. Every element earns its place, with no redundant information. It could be slightly more front-loaded by specifying the action verb, but overall it's appropriately concise.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no annotations, no output schema, and 2 parameters, the description is incomplete. It doesn't explain what format the intelligence comes in (e.g., reports, raw data, summaries), how results are structured, whether there are limitations, or what the agent can expect after invocation. For a tool with multiple siblings and no structured behavioral hints, more context is needed.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, with both parameters ('region' and 'sector') clearly documented in the schema. The description adds no parameter-specific information beyond what the schema provides. According to scoring rules, when schema coverage is high (>80%), the baseline is 3 even with no param info in the description.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool provides intelligence on UK public procurement, listing specific data types (tenders, awarded contracts, buyer profiles, average values) and sources (Contracts Finder, Find a Tender). It distinguishes from siblings by focusing on procurement rather than education, energy, legal, or other domains. However, it doesn't specify the exact verb (e.g., 'search' or 'retrieve'), keeping it at 4 instead of 5.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus the 13 sibling tools. It mentions data sources but doesn't explain how procurement intelligence differs from market sizing, trade intelligence, or other related tools. There's no mention of prerequisites, alternatives, or exclusions, leaving the agent with minimal context for selection.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
uk_trade_intelligenceCRead-onlyIdempotentInspect
Look up UK import duty rates, preferential trade agreement rates, and tariff quotas for any commodity. Accepts an HS commodity code (e.g. '0201' for beef) or a plain-text product description matched to the nearest tariff heading. Returns MFN duty rate, preferential rates by country or trade agreement, quota counts, total trade measures, and a proprietary Tariff Impact Score (0-100, rated FAVOURABLE/MODERATE/COMPLEX/RESTRICTIVE/PROHIBITIVE) assessing overall tariff burden. Use this tool for landed cost calculations, trade compliance research, or post-Brexit tariff comparisons. This tool covers customs duties only; for government procurement contracts, use uk_tenders_intelligence instead. Source: HMRC Trade Tariff API.
| Name | Required | Description | Default |
|---|---|---|---|
| depth | No | Controls response detail. summary: commodity match, MFN duty rate, and VAT only. standard (default): adds preferential rates (up to 10 countries), quota count, and total measures. full: adds all preferential rates without limit. | standard |
| commodity | Yes | HS commodity code (e.g. "0201" for beef, "8471" for computers) or plain-text product description (e.g. "fresh strawberries"). Codes are matched exactly; text triggers a search. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden for behavioral disclosure. It describes what data is returned (duty rates, preferential rates, quotas, trade volumes) but doesn't cover important behavioral aspects like whether this is a read-only operation, potential rate limits, authentication requirements, error conditions, or response format. The description provides basic functionality but lacks operational context.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is appropriately concise with two clear sentences. The first sentence efficiently lists the intelligence types provided, and the second sentence specifies the data source. There's no unnecessary verbiage, and information is front-loaded. It could potentially be slightly more structured but is generally efficient.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a single-parameter tool with no output schema and no annotations, the description provides adequate but minimal context. It explains what data the tool returns and the source, which covers basic functionality. However, it lacks information about response format, error handling, and operational constraints that would be helpful for an AI agent. The description meets minimum viable standards but has clear gaps.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100% for the single parameter 'commodity', which is well-documented in the schema. The description doesn't add any additional parameter semantics beyond what's in the schema. It doesn't provide examples of valid commodity codes beyond the schema's example, nor does it explain the relationship between code and description inputs. Baseline 3 is appropriate when schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool provides UK trade and customs intelligence including duty rates, preferential rates, quotas, and trade volumes for commodities, with a specific source (HMRC Trade Tariff). It distinguishes from siblings by focusing on trade/customs data rather than education, energy, legal, or other intelligence domains. However, it doesn't explicitly contrast with all siblings beyond the trade focus.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. It doesn't mention when this tool is appropriate, when other tools might be better suited, or any prerequisites for usage. The context is implied through the title but not explicitly stated.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
uk_transport_intelligenceCRead-onlyIdempotentInspect
Assess physical transport connectivity for any UK postcode. Returns nearest railway stations with walking distance and train operating companies, bus stops with route numbers and operators, road traffic flow counts, and a composite Connectivity Score. Use this tool for commuter access evaluation, logistics planning, or comparing transport links between locations. For digital connectivity (broadband, mobile coverage), use uk_connectivity_intelligence instead. Sources: DfT NaPTAN, DfT Traffic Counts.
| Name | Required | Description | Default |
|---|---|---|---|
| depth | No | Controls response detail. summary: nearest railway stations only. standard (default): adds bus stops, route information, and connectivity score. full: adds road traffic flow data and detailed operator information. | standard |
| postcode | Yes | Full UK postcode (e.g. "SW1A 1AA"). Transport data is returned for the surrounding area. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. While it mentions the data source (DfT NaPTAN), it doesn't describe response format, rate limits, authentication requirements, error handling, or whether this is a read-only operation. For a tool with no annotation coverage, this leaves significant behavioral gaps.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is appropriately concise - a single sentence that packs in the tool's purpose, key capabilities, and data source. It's front-loaded with the core functionality. However, it could be slightly more structured by separating the different intelligence types for better readability.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a tool with 2 parameters (100% schema coverage) but no annotations and no output schema, the description provides adequate basic context about what intelligence is available. However, it doesn't compensate for the lack of output schema by describing what the return values look like or the format of the intelligence data, which leaves gaps in understanding the tool's complete behavior.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents both parameters thoroughly. The description doesn't add any parameter-specific information beyond what's in the schema - it doesn't explain the practical implications of choosing different depth levels or postcode formatting requirements. Baseline 3 is appropriate when the schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states what the tool does: provides intelligence on nearest stations, bus stops, traffic flow, and connectivity score for UK postcodes. It specifies the source (DfT NaPTAN) and distinguishes from siblings by focusing on transport/connectivity rather than education, health, property, etc. However, it doesn't explicitly contrast with a specific transport alternative tool.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives. The description doesn't mention any prerequisites, limitations, or comparison with other tools that might provide similar transport data. It simply states what the tool does without contextual usage advice.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
uk_vat_validationARead-onlyIdempotentInspect
Validate a UK VAT registration number against HMRC's official register. Returns whether the VAT number is valid, the registered company name, and registered business address. Use this tool to verify a supplier or customer's VAT registration, confirm trading names match, or check VAT status before issuing invoices. For full company due diligence (officers, filings, distress score), use uk_entity_intelligence instead. Source: HMRC VAT Check API.
| Name | Required | Description | Default |
|---|---|---|---|
| vat_number | Yes | UK VAT registration number — 9 digits, optionally prefixed with "GB" (e.g., "123456789" or "GB123456789"). Spaces and hyphens are stripped automatically. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Annotations already declare readOnlyHint=true, idempotentHint=true, and destructiveHint=false. The description adds context about the official register and the return fields (validity, name, address), which goes beyond the annotations without contradiction.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two well-structured sentences deliver purpose, usage, alternative, and source with zero waste. Every word serves a purpose.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
With a single parameter, full schema coverage, no output schema, and annotations covering safety, the description provides all necessary context: purpose, usage guidance, alternative, return fields, and data source.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema covers vat_number with 100% description coverage. The description adds the note about automatic stripping of spaces and hyphens, which is useful behavioral detail beyond the schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description starts with a clear verb+resource ('Validate a UK VAT registration number') and specifies the source (HMRC's official register). It explicitly distinguishes from a sibling tool (uk_entity_intelligence) by stating when to use the alternative.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Provides explicit use cases ('verify a supplier or customer's VAT registration, confirm trading names match, or check VAT status before issuing invoices') and a clear when-not-to-use statement with a named alternative (uk_entity_intelligence).
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
uk_vehicle_intelligenceCRead-onlyIdempotentInspect
Retrieve MOT test history for any UK-registered vehicle. Returns test results (pass/fail/refused), advisory notices, failure items, recorded mileage at each test, a structured Vehicle Health Assessment (0-100, rated EXCELLENT/GOOD/FAIR/POOR/CRITICAL) with component scoring, and enhanced mileage anomaly detection with severity grading (none/minor/major). Use this tool to check a used vehicle before purchase, verify mileage consistency, or assess fleet condition. Returns an error if the registration is not found or has no MOT history (e.g. vehicles under 3 years old). Source: DVSA MOT History API.
| Name | Required | Description | Default |
|---|---|---|---|
| depth | No | Controls response detail. summary: vehicle profile, latest MOT test, and health assessment only. standard (default): last 10 MOT tests with mileage trend. full: complete MOT history (all tests) with mileage anomaly detection and severity grading (none/minor/major). | standard |
| registration | Yes | UK vehicle registration number (e.g. "AB12CDE" or "AB12 CDE"). Case insensitive, spaces are ignored. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden for behavioral disclosure. It lists data types returned (MOT history, mileage trend, etc.) but doesn't mention critical behaviors like whether this is a read-only operation, rate limits, authentication needs, error handling, or data freshness. For a tool with no annotation coverage, this leaves significant gaps in understanding how it behaves.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is efficiently structured in a single sentence that front-loads key data points (MOT history, mileage trend, etc.) and includes the source. There's no wasted text, but it could be slightly more concise by integrating the source note more seamlessly. Every element serves a purpose, making it appropriately sized.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given one parameter with full schema coverage and no output schema, the description is moderately complete. It outlines the types of data returned, which helps the agent understand the scope, but lacks details on output format, error cases, or behavioral constraints. For a simple lookup tool, this is adequate but leaves room for improvement in guiding the agent on what to expect.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, with the single parameter 'registration' clearly documented in the schema as a UK vehicle registration number. The description adds no additional parameter semantics beyond what the schema provides, such as format examples or validation rules. With high schema coverage, the baseline score of 3 is appropriate as the description doesn't compensate but doesn't need to.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool retrieves UK vehicle intelligence data including MOT history, mileage trend, advisories, failures, and Vehicle Health Score. It specifies the source (DVSA MOT History) and distinguishes from siblings by focusing on vehicle data rather than education, property, or other intelligence types. However, it doesn't explicitly name the verb (e.g., 'retrieve' or 'get'), slightly reducing specificity.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives. While it implicitly suggests usage for vehicle-related intelligence, there's no mention of when not to use it, prerequisites, or comparisons with sibling tools like 'uk_transport_intelligence' that might overlap. The agent must infer usage context from the title alone.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
uk_water_intelligenceARead-onlyIdempotentInspect
Assess water quality and pollution risk for any UK postcode. Returns storm overflow discharges (spill frequency, duration, receiving waterway), bathing water classifications, waterbody ecological and chemical status under the Water Framework Directive, water company Ofwat performance metrics, and a Water Quality Score (0-100). Use this tool for environmental due diligence near rivers or coastline, pollution exposure assessment, or water company research. For broader environmental risks (flood zones, geology, radon), use uk_environmental_risk instead. Sources: EA Water Quality Archive, EA EDM Storm Overflows, EA Bathing Water Classifications, Ofwat.
| Name | Required | Description | Default |
|---|---|---|---|
| depth | No | Controls response detail. summary: Water Framework Directive status and storm overflow count only. standard (default): adds bathing water classifications, water company performance, and Water Quality Score. full: adds chemical status detail, full spill history with durations, and receiving waterway information. | standard |
| postcode | Yes | Full UK postcode (e.g. "SW1A 1AA"). Water quality data is returned for the catchment area around this postcode. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, but the description explicitly states input requirements (postcode) and output scope (water quality scores and data categories). It does not mention API limits or error cases, but given the absence of annotations, it provides sufficient behavioral detail.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is concise, front-loaded with the tool's purpose, and each sentence adds unique value: first sentence defines scope, second sentence lists sources. No redundancy.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool has no output schema and two parameters, the description adequately explains input and output context. The depth parameter is well-documented with levels, though the return format is not specified. Still complete enough for correct invocation.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema covers both parameters with full descriptions. The description adds meaning by mapping depth levels to output components ('summary = WFD + overflow count', etc.), enhancing understanding beyond the schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description explicitly states the tool provides 'UK Water Quality Intelligence' and lists specific data types (storm overflow discharges, bathing water classifications, etc.) and sources, clearly distinguishing it from sibling tools that focus on connectivity, demographics, or other domains.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The tool's purpose is clear for water quality queries, but no explicit guidance is given on when to prefer it over sibling tools or when not to use it. However, the sibling names indicate focus areas, and the description's specification of water-related data implicitly guides usage.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
Claim this connector by publishing a /.well-known/glama.json file on your server's domain with the following structure:
{
"$schema": "https://glama.ai/mcp/schemas/connector.json",
"maintainers": [{ "email": "your-email@example.com" }]
}The email address must match the email associated with your Glama account. Once published, Glama will automatically detect and verify the file within a few minutes.
Control your server's listing on Glama, including description and metadata
Access analytics and receive server usage reports
Get monitoring and health status updates for your server
Feature your server to boost visibility and reach more users
For users:
Full audit trail – every tool call is logged with inputs and outputs for compliance and debugging
Granular tool control – enable or disable individual tools per connector to limit what your AI agents can do
Centralized credential management – store and rotate API keys and OAuth tokens in one place
Change alerts – get notified when a connector changes its schema, adds or removes tools, or updates tool definitions, so nothing breaks silently
For server owners:
Proven adoption – public usage metrics on your listing show real-world traction and build trust with prospective users
Tool-level analytics – see which tools are being used most, helping you prioritize development and documentation
Direct user feedback – users can report issues and suggest improvements through the listing, giving you a channel you would not have otherwise
The connector status is unhealthy when Glama is unable to successfully connect to the server. This can happen for several reasons:
The server is experiencing an outage
The URL of the server is wrong
Credentials required to access the server are missing or invalid
If you are the owner of this MCP connector and would like to make modifications to the listing, including providing test credentials for accessing the server, please contact support@glama.ai.
Discussions
No comments yet. Be the first to start the discussion!