guif
Server Details
GBIF MCP — wraps the Global Biodiversity Information Facility API v1 (free, no auth)
- Status
- Healthy
- Last Tested
- Transport
- Streamable HTTP
- URL
- Repository
- pipeworx-io/mcp-gbif
- GitHub Stars
- 0
Glama MCP Gateway
Connect through Glama MCP Gateway for full control over tool access and complete visibility into every call.
Full call logging
Every tool call is logged with complete inputs and outputs, so you can debug issues and audit what your agents are doing.
Tool access control
Enable or disable individual tools per connector, so you decide what your agents can and cannot do.
Managed credentials
Glama handles OAuth flows, token storage, and automatic rotation, so credentials never expire on your clients.
Usage analytics
See which tools your agents call, how often, and when, so you can understand usage patterns and catch anomalies.
Tool Definition Quality
Average 3.7/5 across 3 of 3 tools scored. Lowest: 3.1/5.
Each tool has a clearly distinct purpose: get_occurrences retrieves occurrence records, get_species provides taxonomic details for a known taxon key, and search_species finds species by name or keyword. There is no overlap in functionality, making it easy for an agent to select the right tool.
All tool names follow a consistent verb_noun pattern with snake_case: get_occurrences, get_species, and search_species. This uniformity enhances readability and predictability, making the tool set easy to navigate.
With only 3 tools, the server feels thin for a GBIF biodiversity data domain. While the tools cover key operations (search, retrieve details, get occurrences), typical biodiversity workflows might require additional tools like filtering occurrences by date or region, or managing datasets.
The tool set covers core GBIF operations: searching for species, getting taxonomic details, and retrieving occurrence records. However, there are minor gaps, such as no tools for updating or deleting data, or for more advanced filtering beyond country codes, which agents might need for complex queries.
Available Tools
3 toolsget_occurrencesAInspect
Retrieve georeferenced occurrence records for a taxon. Optionally filter by ISO 3166-1 alpha-2 country code.
| Name | Required | Description | Default |
|---|---|---|---|
| key | Yes | GBIF taxon key (integer) | |
| limit | No | Maximum records to return (1-300, default 20) | |
| country | No | ISO 3166-1 alpha-2 country code to filter occurrences (e.g., "US", "DE") |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It discloses the tool's behavior as a retrieval operation with optional filtering, but it lacks details on rate limits, authentication needs, error handling, or response format. For a tool with no annotations, this is a moderate gap in behavioral context.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is front-loaded with the core purpose in the first sentence and adds a concise optional feature in the second. It uses no wasted words, is appropriately sized for the tool's complexity, and every sentence earns its place by providing essential information.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (3 parameters, no output schema, no annotations), the description is adequate but incomplete. It covers the purpose and basic usage but lacks details on behavioral traits like response structure or limitations. Without annotations or output schema, more context would be helpful for full completeness.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the input schema fully documents the parameters (key, limit, country). The description adds minimal value beyond the schema by mentioning the optional country filter, but it does not provide additional syntax or format details. Baseline 3 is appropriate as the schema handles most of the parameter documentation.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the verb ('Retrieve') and resource ('georeferenced occurrence records for a taxon'), specifying the data type and scope. It distinguishes from sibling tools like 'get_species' and 'search_species' by focusing on occurrence records rather than species information, making the purpose specific and differentiated.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides clear context for when to use the tool ('Retrieve georeferenced occurrence records for a taxon') and includes an optional filter ('Optionally filter by ISO 3166-1 alpha-2 country code'), but it does not explicitly state when not to use it or name alternatives like the sibling tools, leaving some guidance implicit rather than fully explicit.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
get_speciesAInspect
Get full taxonomic details for a GBIF species by its integer taxon key. Use search_species first to find the key.
| Name | Required | Description | Default |
|---|---|---|---|
| key | Yes | GBIF taxon key (integer) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It mentions the tool retrieves 'full taxonomic details,' which implies a read-only operation, but lacks details on rate limits, error handling, or response format. The description adds some behavioral context but is incomplete for a tool with no annotations.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is two sentences, front-loaded with the main purpose and followed by usage guidance. Every sentence earns its place with no wasted words, making it highly efficient and well-structured.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's low complexity (1 parameter, no output schema, no annotations), the description is mostly complete. It covers purpose and usage well but lacks details on behavioral aspects like response format or error handling, which would be beneficial for full completeness.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema fully documents the 'key' parameter. The description adds minimal value by specifying 'integer taxon key,' which is already in the schema. Baseline 3 is appropriate as the schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with specific verbs ('Get full taxonomic details') and resource ('GBIF species'), and distinguishes it from sibling tools by mentioning 'search_species' for finding keys. It precisely communicates what the tool does without being vague or tautological.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides explicit guidance on when to use this tool ('by its integer taxon key') and when to use an alternative ('Use search_species first to find the key'). It clearly differentiates this tool from its sibling, offering complete usage context.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
search_speciesBInspect
Search GBIF species backbone by name or keyword. Returns matched taxa with rank, status, and classification.
| Name | Required | Description | Default |
|---|---|---|---|
| limit | No | Maximum results to return (1-100, default 20) | |
| query | Yes | Species name or keyword (e.g., "Homo sapiens", "oak") |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions what the tool returns ('matched taxa with rank, status, and classification'), which is helpful, but lacks details on error handling, rate limits, authentication needs, or whether this is a read-only operation. For a search tool with zero annotation coverage, this leaves significant gaps in understanding its behavior.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, well-structured sentence that efficiently conveys the tool's purpose, search method, and return values. It is front-loaded with key information and avoids unnecessary words, making it easy for an agent to parse quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's moderate complexity (search with two parameters) and lack of annotations or output schema, the description is minimally adequate. It covers the basic purpose and return format but omits important contextual details like error conditions, pagination, or how results are ordered. Without an output schema, the agent must rely on the description's brief mention of return values, which is insufficient for full understanding.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 100% description coverage, clearly documenting both parameters ('query' and 'limit') with their purposes and constraints. The description adds minimal value beyond the schema, only implying the search scope ('GBIF species backbone') without providing additional syntax or format details. This meets the baseline score of 3 when the schema does the heavy lifting.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose with a specific verb ('Search') and resource ('GBIF species backbone'), and mentions the search criteria ('by name or keyword'). It doesn't explicitly differentiate from sibling tools like 'get_species', but the search focus is clear. The description avoids tautology by providing meaningful context beyond just the tool name.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives like 'get_species' or 'get_occurrences'. It mentions the search functionality but doesn't specify scenarios where this tool is preferred or excluded, leaving the agent to infer usage based on the tool name alone.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
Claim this connector by publishing a /.well-known/glama.json file on your server's domain with the following structure:
{
"$schema": "https://glama.ai/mcp/schemas/connector.json",
"maintainers": [{ "email": "your-email@example.com" }]
}The email address must match the email associated with your Glama account. Once published, Glama will automatically detect and verify the file within a few minutes.
Control your server's listing on Glama, including description and metadata
Access analytics and receive server usage reports
Get monitoring and health status updates for your server
Feature your server to boost visibility and reach more users
For users:
Full audit trail – every tool call is logged with inputs and outputs for compliance and debugging
Granular tool control – enable or disable individual tools per connector to limit what your AI agents can do
Centralized credential management – store and rotate API keys and OAuth tokens in one place
Change alerts – get notified when a connector changes its schema, adds or removes tools, or updates tool definitions, so nothing breaks silently
For server owners:
Proven adoption – public usage metrics on your listing show real-world traction and build trust with prospective users
Tool-level analytics – see which tools are being used most, helping you prioritize development and documentation
Direct user feedback – users can report issues and suggest improvements through the listing, giving you a channel you would not have otherwise
The connector status is unhealthy when Glama is unable to successfully connect to the server. This can happen for several reasons:
The server is experiencing an outage
The URL of the server is wrong
Credentials required to access the server are missing or invalid
If you are the owner of this MCP connector and would like to make modifications to the listing, including providing test credentials for accessing the server, please contact support@glama.ai.
Discussions
No comments yet. Be the first to start the discussion!