Skip to main content
Glama

Server Details

AccessOracle - 10 access control tools: IAM, PAM, recertification, segregation of duties.

Status
Healthy
Last Tested
Transport
Streamable HTTP
URL
Repository
ToolOracle/accessoracle
GitHub Stars
0

Glama MCP Gateway

Connect through Glama MCP Gateway for full control over tool access and complete visibility into every call.

MCP client
Glama
MCP server

Full call logging

Every tool call is logged with complete inputs and outputs, so you can debug issues and audit what your agents are doing.

Tool access control

Enable or disable individual tools per connector, so you decide what your agents can and cannot do.

Managed credentials

Glama handles OAuth flows, token storage, and automatic rotation, so credentials never expire on your clients.

Usage analytics

See which tools your agents call, how often, and when, so you can understand usage patterns and catch anomalies.

100% free. Your data is private.
Tool DescriptionsC

Average 2.7/5 across 10 of 10 tools scored. Lowest: 2/5.

Server CoherenceA
Disambiguation5/5

Each tool targets a distinct IAM/compliance function: gap analysis, recertification reviews, emergency logging, health check, JML process, account listing, MFA compliance, privileged audit, account registration, and SoD conflicts. No overlap in purposes.

Naming Consistency4/5

All names use underscores and are descriptive, but patterns vary: some are verb_noun (list_accounts, register_account, break_glass_log) while others are noun_phrase (access_gap_analysis, mfa_compliance, sod_matrix). Consistent in style but not fully uniform.

Tool Count5/5

10 tools is well-scoped for an access governance server, covering core functions without being overwhelming. Each tool earns its place.

Completeness4/5

Covers essential IAM operations (account listing/registration, compliance checks, logging, review tracking). Missing update/delete for accounts and policy management, but these are less critical for an analysis-focused server.

Available Tools

10 tools
access_gap_analysisCInspect

Gap analysis against RTS Art. 21 requirements.

ParametersJSON Schema
NameRequiredDescriptionDefault

No parameters

Behavior1/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description must convey behavioral traits, but it only states the tool performs a gap analysis. It does not indicate whether this is a read-only operation, what side effects exist, or whether authentication or specific permissions are required.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single sentence of 8 words, very concise and front-loaded. It could arguably be expanded with more details, but it does not waste words.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given no parameters and no output schema, the tool is simple, but the description fails to explain what the gap analysis produces (e.g., a report, list, or score) or any context about RTS Art. 21. This omission reduces completeness.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

There are zero parameters, and schema description coverage is 100%. According to the rules, the baseline for 0 parameters is 4, and the description does not need to add parameter information. It is adequate.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description specifies the verb 'gap analysis' and the resource 'RTS Art. 21 requirements', making the purpose clear. However, it does not distinguish this tool from siblings like access_review or sod_matrix, so it misses the top score.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance is provided on when to use this tool versus alternatives. There is no mention of prerequisites or typical use cases, leaving the agent to infer from the name alone.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

access_reviewCInspect

Track access recertification reviews. Set add=true to log a review.

ParametersJSON Schema
NameRequiredDescriptionDefault
addNo
notesNo
decisionNo
reviewerNo
account_idNo
review_dateNo
overdue_daysNo
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description must convey behavioral traits. It only mentions logging a review, but does not disclose side effects, permissions required, data persistence, or what happens when 'add' is false. This is insufficient for an agent to understand the tool's behavior.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness2/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is extremely short (one sentence), but conciseness should not come at the cost of completeness. It under-specifies the tool's functionality and parameters, making it inadequate despite being brief.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness1/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given 7 parameters, no output schema, and no annotations, the description is woefully incomplete. It barely introduces the tool's purpose and fails to provide essential details for correct invocation and understanding.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters1/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema has 7 parameters with 0% description coverage. The description only addresses the 'add' parameter, ignoring the meanings or usage of 'notes', 'decision', 'reviewer', 'account_id', 'review_date', and 'overdue_days'. The agent cannot infer their semantics from the given description.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose3/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description states 'Track access recertification reviews' which gives a general purpose but lacks specificity. Mentioning 'Set add=true to log a review' adds some action context, but it does not clearly distinguish the tool from siblings like 'access_gap_analysis' or 'break_glass_log'.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No explicit guidance on when to use this tool versus alternatives. The only usage instruction is 'Set add=true to log a review', which is minimal and does not address prerequisites or exclusion cases.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

break_glass_logCInspect

Emergency access logging. Set log=true to record usage.

ParametersJSON Schema
NameRequiredDescriptionDefault
logNo
reasonNo
used_byNo
account_idNo
approved_byNo
duration_minutesNo
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description must fully disclose behavior. It only mentions that setting log=true records usage, leaving return values, side effects, and parameter interactions undocumented.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness2/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is very concise (two sentences) but at the expense of providing necessary details for a 6-parameter tool. It is under-specified rather than appropriately brief.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given 6 parameters, no output schema, and no annotations, the description is incomplete. It fails to explain what the tool returns or how parameters interact, leaving an agent without sufficient context.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters2/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 0%. The description only addresses the 'log' parameter, ignoring the other five (reason, used_by, account_id, approved_by, duration_minutes) that remain undefined.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description states 'Emergency access logging' which clearly identifies the domain and action. It is more specific than a generic 'log' tool, though it lacks detail on what exactly logging entails.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance on when to use this tool versus alternatives like access_review. There are no exclusions or context provided.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

health_checkCInspect

Server status.

ParametersJSON Schema
NameRequiredDescriptionDefault

No parameters

Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description carries full burden but only says 'Server status'. It does not disclose whether this is a read operation, if it has side effects, or what the response looks like.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Extremely concise at two words, which is appropriate for a simple health check. However, it lacks any structure or additional context that could aid understanding.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a tool with no parameters and no output schema, the description is the only source of behavior. 'Server status' is insufficient to distinguish from other tools or to set expectations about the output.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

No parameters exist, so schema coverage is 100%. The description adds no parameter-level info, but baseline 3 applies since schema already covers all params.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose3/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description 'Server status' indicates a health check, but lacks specificity on what status entails (e.g., connectivity, resource usage). It is not misleading but is vague, providing only a generic verb-resource pairing.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance on when to use this tool vs siblings like 'access_gap_analysis' or 'break_glass_log'. The agent has no context to decide based on task requirements.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

jml_processCInspect

Joiner/Mover/Leaver process tracking. Set log_event=true to log.

ParametersJSON Schema
NameRequiredDescriptionDefault
dateNo
usernameNo
log_eventNo
event_typeNo
verified_byNo
actions_takenNo
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are present, so the description carries full responsibility for behavioral disclosure. It only mentions that setting log_event=true enables logging, but does not state whether the tool is read-only, modifies state, or has any side effects. The nature of 'process tracking' implies mutation, but this is not clarified.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness2/5

Is the description appropriately sized, front-loaded, and free of redundancy?

At only two sentences, the description is minimal, but it lacks substance rather than being efficiently concise. It fails to earn its length by omitting critical details such as the primary action or parameter semantics, making it under-specified rather than concise.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness1/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's complexity (6 parameters, no output schema, no annotations), the description is severely incomplete. It does not explain the overall workflow, expected output, or prerequisites, leaving the agent with insufficient information to use the tool correctly.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters2/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

With 0% schema description coverage across 6 parameters, the description must compensate but only addresses one parameter (log_event). It clarifies that log_event controls logging, but provides no meaning for date, username, event_type, verified_by, or actions_taken, leaving the agent without guidance on their purpose.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose3/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description states 'Joiner/Mover/Leaver process tracking', which identifies the resource (JML events) and implies a tracking/recording function. However, it lacks a clear verb (e.g., 'create', 'list', or 'update') to specify the exact action, leaving ambiguity whether this tool logs, queries, or processes JML records.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance is provided on when to use this tool versus its siblings (e.g., access_gap_analysis, break_glass_log). The description does not create any context for selection or exclusion, leaving the agent to guess the appropriate scenario.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

list_accountsCInspect

List accounts with filters.

ParametersJSON Schema
NameRequiredDescriptionDefault
no_mfaNo
systemNo
account_typeNo
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description carries full burden for behavioral disclosure. It only says 'with filters,' omitting whether this is a read operation, any side effects, authentication needs, or pagination. A read operation is implied, but not stated explicitly.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness3/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is extremely short (4 words), which is concise but overly minimal. It fronts the purpose but sacrifices detail. A bit more structure or elaboration would be beneficial without losing conciseness.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness1/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given 3 optional parameters, no output schema, and no annotations, the description is severely incomplete. It does not explain filter behavior, return format, or how this tool fits with siblings. Agents would struggle to use it correctly without additional context.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters2/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 0%, meaning the description must explain parameters. It only mentions 'filters' generically without describing the three specific parameters (no_mfa, system, account_type). The parameter names provide some self-documentation, but the description adds no additional meaning.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description states 'List accounts with filters,' which is a clear verb+resource combination. However, 'accounts' is vague and doesn't specify what kind of accounts, nor does it differentiate from sibling tools like access_review or mfa_compliance. But it still conveys the basic action.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance on when to use this tool vs alternatives. The description does not mention prerequisites, when not to use, or provide any context about appropriate scenarios. Sibling tools exist but are not referenced.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

mfa_complianceAInspect

Check MFA coverage against DORA Art. 9(4)(d) requirements.

ParametersJSON Schema
NameRequiredDescriptionDefault

No parameters

Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description carries the full burden. It implies a read-only check but does not explicitly state side effects, authentication needs, or any behavioral constraints beyond the basic purpose.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

A single sentence that is front-loaded with the action and resource, efficient and to the point.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given zero parameters and no output schema, the description is mostly complete for a simple check tool. However, it could briefly mention what the output looks like (e.g., compliance status) to reduce ambiguity.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters5/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The input schema has zero parameters, so the schema provides no meaning. The description adds value by specifying the regulatory context (DORA), which is essential for correct usage.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description uses a specific verb ('Check') and resource ('MFA coverage'), and adds regulatory context ('DORA Art. 9(4)(d)'), clearly distinguishing it from sibling tools like access_gap_analysis or access_review.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance is provided on when to use this tool versus alternatives, nor any prerequisites or exclusions.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

privileged_auditBInspect

Privileged account audit — MFA, shared, review status.

ParametersJSON Schema
NameRequiredDescriptionDefault

No parameters

Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, and the description does not disclose behavioral traits such as whether the tool is read-only, requires admin privileges, or has any side effects. The agent cannot infer safety or permissions from this description.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is extremely concise, fitting into a single line with three key terms. It is front-loaded and contains no unnecessary words, though it may be too terse for an agent to fully grasp the tool's purpose.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given zero parameters and no output schema, the description is minimally adequate. However, it lacks context on how this tool differs from similar siblings like mfa_compliance or access_review, and does not specify the output format or expected results.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The input schema has zero parameters, so schema coverage is 100%. The description adds useful context about the audit scope (MFA, shared, review status), which helps the agent understand what the tool does despite no parameters.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description specifies a clear verb (audit) and resource (privileged accounts) with focus areas (MFA, shared, review status). However, it does not distinguish itself from sibling tools like mfa_compliance or access_review, which may overlap in functionality.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance on when to use this tool versus alternatives. For example, it is unclear whether to use privileged_audit or mfa_compliance for checking MFA status, or access_review for review status.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

register_accountCInspect

Register a privileged/service/shared account for IAM tracking.

ParametersJSON Schema
NameRequiredDescriptionDefault
notesNo
ownerNo
systemNo
usernameYes
is_sharedNo
account_idNo
cif_accessNo
departmentNo
mfa_methodNo
criticalityNo
mfa_enabledNo
account_typeYes
remote_accessNo
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations; description only says 'Register' without detailing side effects, duplicate handling, required permissions, or output. Lacks behavioral disclosure expected for a creation tool.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness3/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Single sentence is concise, but it misleads by implying only privileged/service/shared account types, while schema includes 'standard' and 'break_glass'. Accuracy issue reduces effectiveness.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

With 13 parameters, no output schema, and no annotations, the description is too brief. It fails to explain required parameters, constraints, or post-registration behavior, leaving significant gaps.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters1/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema has 0% description coverage, and the description adds no parameter-level meaning. Agent gets no help understanding what each parameter does beyond the schema types.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the action ('Register') and the resource ('account for IAM tracking'), with specific account types (privileged/service/shared). It distinguishes from sibling tools like list_accounts or access_review.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance on when to use this tool versus alternatives (e.g., jml_process). No prerequisites, exclusions, or context provided.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

sod_matrixCInspect

Detect Segregation of Duties conflicts across accounts.

ParametersJSON Schema
NameRequiredDescriptionDefault

No parameters

Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description carries full responsibility for behavioral disclosure. It does not state whether the tool is read-only, whether it requires authentication, or what happens on execution. The lack of detail limits transparency.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single sentence with no unnecessary words, achieving high efficiency. However, it could be expanded to include more useful context without becoming verbose.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given no output schema or annotations, the description should provide more context about return values, behavior, and use cases. It only states the basic purpose, which is insufficient for complete understanding.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The input schema has zero parameters, and schema coverage is 100% (by default). The description adds no parameter information, but no additional meaning is needed beyond the schema. Baseline of 3 is appropriate.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool detects Segregation of Duties conflicts across accounts. It uses a specific verb ('detect') and specifies the resource ('SoD conflicts'). However, it does not differentiate from sibling tools like access_gap_analysis or access_review, which may overlap in functionality.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance is provided on when to use this tool versus its siblings. The description does not specify prerequisites, exclusions, or alternative tools, leaving the agent without context for selection.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

Discussions

No comments yet. Be the first to start the discussion!

Try in Browser

Your Connectors

Sign in to create a connector for this server.