Skip to main content
Glama

Server Details

Twilio MCP Pack — send SMS, list messages, make calls via Twilio REST API.

Status
Healthy
Last Tested
Transport
Streamable HTTP
URL
Repository
pipeworx-io/mcp-twilio
GitHub Stars
0

Glama MCP Gateway

Connect through Glama MCP Gateway for full control over tool access and complete visibility into every call.

MCP client
Glama
MCP server

Full call logging

Every tool call is logged with complete inputs and outputs, so you can debug issues and audit what your agents are doing.

Tool access control

Enable or disable individual tools per connector, so you decide what your agents can and cannot do.

Managed credentials

Glama handles OAuth flows, token storage, and automatic rotation, so credentials never expire on your clients.

Usage analytics

See which tools your agents call, how often, and when, so you can understand usage patterns and catch anomalies.

100% free. Your data is private.
Tool DescriptionsA

Average 3.7/5 across 9 of 10 tools scored.

Server CoherenceA
Disambiguation3/5

The Twilio-specific tools (twilio_send_sms, twilio_get_message, twilio_list_messages, twilio_make_call, twilio_list_calls) are clearly distinct. However, the Pipeworx tools (ask_pipeworx, discover_tools, remember, recall, forget) have overlapping purposes; for example, ask_pipeworx and discover_tools both involve finding information, and remember/recall/forget form a memory system that could be confused with session context.

Naming Consistency4/5

Twilio tools use a consistent 'twilio_' prefix with verb_noun pattern (e.g., twilio_send_sms, twilio_list_calls). The Pipeworx tools also follow verb_noun style (ask_pipeworx, discover_tools, remember, recall, forget). The two groups have different prefixes, which is acceptable given they belong to distinct subsystems.

Tool Count4/5

With 10 tools, the count is reasonable for a server that combines Twilio messaging/voice capabilities and a general-purpose query/memory system. The set feels slightly broad but not overloaded.

Completeness3/5

Twilio tools cover sending SMS, making calls, listing messages and calls, but lack operations like deleting messages, updating calls, or managing phone numbers. The Pipeworx side offers querying, tool discovery, and memory, but there is no obvious gap given its open-ended nature.

Available Tools

10 tools
ask_pipeworxAInspect

Ask a question in plain English and get an answer from the best available data source. Pipeworx picks the right tool, fills the arguments, and returns the result. No need to browse tools or learn schemas — just describe what you need. Examples: "What is the US trade deficit with China?", "Look up adverse events for ozempic", "Get Apple's latest 10-K filing".

ParametersJSON Schema
NameRequiredDescriptionDefault
questionYesYour question or request in natural language
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Without annotations, the description carries the full burden. It clearly states that the tool internally picks the right tool and fills arguments, which is a key behavioral trait. However, it doesn't disclose potential latency, error handling, or data source limitations. The 5 is missed because it doesn't mention what happens if no data source is available.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is extremely concise: three sentences plus examples. Every sentence adds value, no filler. Front-loaded with the core purpose, followed by behavioral insight, then usage guidance via examples.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the low complexity (single required parameter, no output schema), the description is mostly complete. It explains what the tool does, how it works, and when to use it. Missing details about output format or error cases, but the examples compensate. A 5 would require a brief note on return format.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100% with a single 'question' parameter described as 'Your question or request in natural language.' The description reinforces this with examples but adds no extra semantic details about parameter constraints or formats, so baseline 3 is appropriate.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool's purpose: answering plain English questions using the best available data source. It emphasizes abstraction over browsing tools, with concrete examples like 'What is the US trade deficit with China?' that differentiate it from sibling tools like discover_tools.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines5/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description explicitly tells the agent when to use this tool ('just describe what you need') and when not to ('No need to browse tools or learn schemas'). It also provides multiple examples covering diverse use cases, guiding selection over more specific tools like twilio_send_sms.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

discover_toolsAInspect

Search the Pipeworx tool catalog by describing what you need. Returns the most relevant tools with names and descriptions. Call this FIRST when you have 500+ tools available and need to find the right ones for your task.

ParametersJSON Schema
NameRequiredDescriptionDefault
limitNoMaximum number of tools to return (default 20, max 50)
queryYesNatural language description of what you want to do (e.g., "analyze housing market trends", "look up FDA drug approvals", "find trade data between countries")
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, but the description discloses the search behavior and result format (names and descriptions). It does not detail potential performance implications or edge cases (e.g., empty results), but the core behavior is clear.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Three sentences, each providing essential information: what the tool does, what it returns, and when to use it. No wasted words.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given no output schema, the description could mention return format details (e.g., ranked list, includes IDs). However, the context of 11 sibling tools and the clear usage guideline make it sufficient for an agent to use correctly. The limit parameter is documented in the schema, so it's fine.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100%, so the schema already documents both parameters well. The description adds value by explaining the query parameter's use as 'natural language description' and provides examples like 'analyze housing market trends', which goes beyond the schema's generic description.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description explicitly states 'Search the Pipeworx tool catalog by describing what you need' and 'Returns the most relevant tools with names and descriptions,' clearly identifying the action and resource. It also distinguishes itself from sibling tools by being a meta-tool for discovery, not a domain-specific tool.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines5/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides explicit guidance: 'Call this FIRST when you have 500+ tools available and need to find the right ones for your task.' This tells the agent exactly when to use it and implies that other tools are for specific tasks after discovery.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

forgetBInspect

Delete a stored memory by key.

ParametersJSON Schema
NameRequiredDescriptionDefault
keyYesMemory key to delete
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description must disclose behavior. It correctly indicates the tool deletes data, but does not mention side effects (e.g., permanent deletion, error handling for non-existent keys) or authorization needs. The description is minimal but not contradictory.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single sentence, 6 words, which is concise and front-loaded. However, it lacks additional context that might be useful (e.g., 'Note: this action cannot be undone').

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a simple deletion tool with one parameter and no output schema, the description is adequate but minimal. It does not explain what happens on success/failure, nor does it mention that deletion is irreversible. Given the tool's simplicity, this is acceptable but not exemplary.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100% with one parameter 'key' described as 'Memory key to delete'. The description does not add beyond the schema, but the schema is sufficient. Since coverage is high, baseline 3 is appropriate; a slight bonus is given for clarity of the parameter purpose.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the action ('Delete') and the resource ('a stored memory by key'). It distinguishes the tool from siblings like 'remember' (create) and 'recall' (read), but does not explicitly differentiate from other deletion-like tools, though none appear to exist.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance is provided on when to use this tool versus alternatives. For instance, it does not mention that the key must exist before deletion or suggest verifying existence with 'recall' first.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

recallAInspect

Retrieve a previously stored memory by key, or list all stored memories (omit key). Use this to retrieve context you saved earlier in the session or in previous sessions.

ParametersJSON Schema
NameRequiredDescriptionDefault
keyNoMemory key to retrieve (omit to list all keys)
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations provided, so description must bear burden. States listing behavior when key omitted, but doesn't detail side effects or persistence details.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two sentences, clear and front-loaded. Every part adds value.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Adequate for a simple retrieval tool with optional parameter. No output schema needed given simple return type implied.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100% and description adds meaning: explains purpose of omitting key (list all) vs specifying it.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

Clearly states it retrieves a memory by key or lists all memories. Distinguishes from 'remember' and 'forget' siblings.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Provides some guidance on when to use (retrieve context), but does not explicitly mention when not to use or alternatives.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

rememberAInspect

Store a key-value pair in your session memory. Use this to save intermediate findings, user preferences, or context across tool calls. Authenticated users get persistent memory; anonymous sessions last 24 hours.

ParametersJSON Schema
NameRequiredDescriptionDefault
keyYesMemory key (e.g., "subject_property", "target_ticker", "user_preference")
valueYesValue to store (any text — findings, addresses, preferences, notes)
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description carries the full burden. It discloses persistence behavior ('Authenticated users get persistent memory; anonymous sessions last 24 hours'), which is useful. However, it does not mention limits on memory size, key uniqueness, or whether overwriting is allowed, leaving some behavioral gaps.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is three sentences, each adding value: first defines the action, second explains usage, third clarifies persistence. No wasted words, front-loaded with the core purpose.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's simplicity (2 string params, no output schema, no annotations), the description is nearly complete. It covers purpose, usage, and persistence. A slight gap is the lack of mention of overwrite behavior or memory limits, but it's sufficient for most agents.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100%, so the baseline is 3. The description adds minimal parameter semantics beyond what the schema provides (e.g., 'key' and 'value' are already described). It does not elaborate on format constraints or usage tips, but the schema is sufficiently clear.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the action ('store'), the resource ('key-value pair in session memory'), and the purpose ('save intermediate findings, user preferences, or context across tool calls'). It distinguishes from siblings like 'forget' and 'recall' by focusing on storage.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description says 'Use this to save intermediate findings, user preferences, or context across tool calls', providing clear usage context. It does not explicitly exclude other tools, but the purpose is specific enough that an agent can infer when to use it (e.g., not for retrieving or deleting).

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

twilio_get_messageBInspect

Get details of a specific Twilio message by its SID.

ParametersJSON Schema
NameRequiredDescriptionDefault
sidYesMessage SID (starts with SM...)
_authTokenYesTwilio Auth Token
_accountSidYesTwilio Account SID
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations are empty, so the description bears full responsibility. It states 'Get details', implying a read-only operation, but does not mention any side effects, rate limits, or required permissions beyond the auth tokens in the schema. Lacks depth but does not contradict any annotations.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, clear sentence that efficiently conveys the tool's purpose without unnecessary words.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the simplicity of the tool (3 parameters, no output schema), the description is adequate. It covers the basic purpose, but could mention that the response contains message details (since no output schema exists) or that the SID is required.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100%, and each parameter has a description in the schema. The tool description adds no additional parameter information beyond what the schema provides, so a baseline score of 3 is appropriate.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the action ('get details') and the resource ('a specific Twilio message by its SID'). It is specific and informative, though it does not explicitly distinguish from sibling tools like twilio_list_messages.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance on when to use this tool versus alternatives. It does not mention that the user needs the message SID or that it is for retrieving a single message, which could be inferred, but no explicit when-to-use or when-not-to-use advice is given.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

twilio_list_callsBInspect

List recent phone calls from your Twilio account. Returns call SID, status, duration, and direction.

ParametersJSON Schema
NameRequiredDescriptionDefault
limitNoMax calls to return (default 20, max 100)
_authTokenYesTwilio Auth Token
_accountSidYesTwilio Account SID
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations are empty, so description carries full burden. It states the tool lists 'recent' calls, implying a time-based default, and mentions returned fields. However, it does not clarify if the tool is read-only, whether authentication parameters are required each call, or any rate limits. The lack of contradiction with annotations (which are empty) means no penalty.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two sentences efficiently convey purpose and return fields. No unnecessary words, but could be slightly more structured (e.g., separate purpose and output).

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a simple list tool with no output schema, the description covers core purpose and output fields. However, it omits details like pagination, default ordering, and whether 'recent' is based on date. Sibling tools and context signals suggest moderate complexity, so a bit more detail would help.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100% with descriptions for all parameters. The description adds no further meaning beyond the schema (e.g., does not explain the 'limit' parameter's effect or how to use account SID/token). Baseline 3 is appropriate since schema already documents parameters.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the action ('List recent phone calls'), resource ('from your Twilio account'), and key return fields ('call SID, status, duration, and direction'). It distinguishes from sibling tools like twilio_make_call and twilio_list_messages by focusing on listing calls.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance is provided on when to use this tool versus alternatives. Sibling tools exist for sending SMS or making calls, but no criteria (e.g., need to filter vs. list all) are mentioned.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

twilio_list_messagesAInspect

List recent SMS/MMS messages from your Twilio account. Supports filtering by to/from number and pagination.

ParametersJSON Schema
NameRequiredDescriptionDefault
toNoFilter by destination phone number
fromNoFilter by sender phone number
limitNoMax messages to return (default 20, max 100)
_authTokenYesTwilio Auth Token
_accountSidYesTwilio Account SID
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description carries the burden. It discloses filtering and pagination behavior but does not mention authentication requirements (though account SID and auth token are required params, not explained in description), rate limits, or default/max limits. The default limit of 20 and max 100 are stated in the schema but not in the description.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single sentence, front-loaded with the main purpose. It is concise and clear, though it could benefit from a brief note on required authentication parameters.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool has 5 parameters, no output schema, and no annotations, the description is somewhat complete but lacks details on authentication (though params indicate), response format, and error handling. It is adequate for a simple listing tool but could be more comprehensive.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, so the schema already documents all parameters. The description adds value by mentioning 'filtering by to/from number' and 'pagination', which maps to the 'to', 'from', and 'limit' parameters, but does not add new meaning beyond the schema.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the verb 'List' and resource 'SMS/MMS messages from your Twilio account', and distinguishes from sibling tools like twilio_send_sms (sending) and twilio_get_message (single message) by specifying 'list' and mentioning filtering and pagination.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description indicates usage for listing messages with filtering by to/from and pagination, but does not explicitly state when not to use it (e.g., for sending messages) or name alternatives directly. However, the sibling tool names provide context.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

twilio_make_callAInspect

Initiate a phone call via Twilio. Requires a TwiML URL or application SID to control call behavior.

ParametersJSON Schema
NameRequiredDescriptionDefault
toYesDestination phone number in E.164 format
urlYesTwiML URL that controls what happens when the call connects
fromYesTwilio phone number to call from in E.164 format
_authTokenYesTwilio Auth Token
_accountSidYesTwilio Account SID
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations provided, so description carries full burden. It discloses the need for TwiML URL and account/auth parameters, but doesn't mention side effects (cost, phone network delays), rate limits, or error conditions.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two sentences, front-loaded with action and requirement. Efficient and clear.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given 5 parameters, all required, and no output schema, description is adequate but could include info about call status responses or how to retrieve call results. Lacks guidance on async nature.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100%, so baseline is 3. Description adds value by explaining that 'url' is a TwiML URL controlling call behavior, and 'from' is a Twilio phone number, which clarifies purpose beyond the schema's basic type descriptions.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Description clearly states the action (initiate a phone call), the resource (via Twilio), and adds key requirement (TwiML URL or application SID). Distinct from sibling tools like twilio_send_sms (SMS) and twilio_list_calls (list).

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Description implies it's for making calls, but does not explicitly state when to use this tool versus alternatives like twilio_send_sms. No guidance on prerequisites or alternatives.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

twilio_send_smsBInspect

Send an SMS message via Twilio. Returns the message SID and status.

ParametersJSON Schema
NameRequiredDescriptionDefault
toYesDestination phone number in E.164 format (e.g., +15551234567)
bodyYesMessage body text (max 1600 characters)
fromYesTwilio phone number to send from in E.164 format
_authTokenYesTwilio Auth Token
_accountSidYesTwilio Account SID (starts with AC...)
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description partially discloses behavior: it sends an SMS and returns the SID and status. However, it does not mention potential side effects (cost, delivery delays), rate limits, or failure modes (e.g., invalid numbers). Annotations are absent, so the description carries full burden and could be more detailed.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is concise at one sentence plus a return value note. It is front-loaded with the action. However, it could be slightly more informative without being verbose.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool has 5 parameters (all required) and no output schema, the description is somewhat complete but lacks context about authentication requirements, character limits, or error handling. The sibling tools list includes twilio_get_message, which could be compared, but no guidance is given.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100%, so the description adds minimal value beyond the schema. It mentions the return value but does not explain parameter details like body character limit or E.164 format beyond what the schema already provides. Baseline 3 is appropriate.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the action ('Send an SMS message'), the resource ('via Twilio'), and the return values ('message SID and status'). It distinguishes from siblings like twilio_make_call and twilio_get_message, but could be more specific about which sibling tools are related.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance is provided on when to use this tool versus alternatives like twilio_list_messages or when not to use it (e.g., for sending to international numbers). The description does not mention prerequisites like having a Twilio account with SMS capabilities.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

Discussions

No comments yet. Be the first to start the discussion!

Try in Browser

Your Connectors

Sign in to create a connector for this server.