Twilio
Server Details
Twilio MCP Pack — send SMS, list messages, make calls via Twilio REST API.
- Status
- Healthy
- Last Tested
- Transport
- Streamable HTTP
- URL
- Repository
- pipeworx-io/mcp-twilio
- GitHub Stars
- 0
Glama MCP Gateway
Connect through Glama MCP Gateway for full control over tool access and complete visibility into every call.
Full call logging
Every tool call is logged with complete inputs and outputs, so you can debug issues and audit what your agents are doing.
Tool access control
Enable or disable individual tools per connector, so you decide what your agents can and cannot do.
Managed credentials
Glama handles OAuth flows, token storage, and automatic rotation, so credentials never expire on your clients.
Usage analytics
See which tools your agents call, how often, and when, so you can understand usage patterns and catch anomalies.
Tool Definition Quality
Average 3.7/5 across 9 of 10 tools scored.
The Twilio-specific tools (twilio_send_sms, twilio_get_message, twilio_list_messages, twilio_make_call, twilio_list_calls) are clearly distinct. However, the Pipeworx tools (ask_pipeworx, discover_tools, remember, recall, forget) have overlapping purposes; for example, ask_pipeworx and discover_tools both involve finding information, and remember/recall/forget form a memory system that could be confused with session context.
Twilio tools use a consistent 'twilio_' prefix with verb_noun pattern (e.g., twilio_send_sms, twilio_list_calls). The Pipeworx tools also follow verb_noun style (ask_pipeworx, discover_tools, remember, recall, forget). The two groups have different prefixes, which is acceptable given they belong to distinct subsystems.
With 10 tools, the count is reasonable for a server that combines Twilio messaging/voice capabilities and a general-purpose query/memory system. The set feels slightly broad but not overloaded.
Twilio tools cover sending SMS, making calls, listing messages and calls, but lack operations like deleting messages, updating calls, or managing phone numbers. The Pipeworx side offers querying, tool discovery, and memory, but there is no obvious gap given its open-ended nature.
Available Tools
10 toolsask_pipeworxAInspect
Ask a question in plain English and get an answer from the best available data source. Pipeworx picks the right tool, fills the arguments, and returns the result. No need to browse tools or learn schemas — just describe what you need. Examples: "What is the US trade deficit with China?", "Look up adverse events for ozempic", "Get Apple's latest 10-K filing".
| Name | Required | Description | Default |
|---|---|---|---|
| question | Yes | Your question or request in natural language |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Without annotations, the description carries the full burden. It clearly states that the tool internally picks the right tool and fills arguments, which is a key behavioral trait. However, it doesn't disclose potential latency, error handling, or data source limitations. The 5 is missed because it doesn't mention what happens if no data source is available.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise: three sentences plus examples. Every sentence adds value, no filler. Front-loaded with the core purpose, followed by behavioral insight, then usage guidance via examples.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the low complexity (single required parameter, no output schema), the description is mostly complete. It explains what the tool does, how it works, and when to use it. Missing details about output format or error cases, but the examples compensate. A 5 would require a brief note on return format.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100% with a single 'question' parameter described as 'Your question or request in natural language.' The description reinforces this with examples but adds no extra semantic details about parameter constraints or formats, so baseline 3 is appropriate.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: answering plain English questions using the best available data source. It emphasizes abstraction over browsing tools, with concrete examples like 'What is the US trade deficit with China?' that differentiate it from sibling tools like discover_tools.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description explicitly tells the agent when to use this tool ('just describe what you need') and when not to ('No need to browse tools or learn schemas'). It also provides multiple examples covering diverse use cases, guiding selection over more specific tools like twilio_send_sms.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
discover_toolsAInspect
Search the Pipeworx tool catalog by describing what you need. Returns the most relevant tools with names and descriptions. Call this FIRST when you have 500+ tools available and need to find the right ones for your task.
| Name | Required | Description | Default |
|---|---|---|---|
| limit | No | Maximum number of tools to return (default 20, max 50) | |
| query | Yes | Natural language description of what you want to do (e.g., "analyze housing market trends", "look up FDA drug approvals", "find trade data between countries") |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, but the description discloses the search behavior and result format (names and descriptions). It does not detail potential performance implications or edge cases (e.g., empty results), but the core behavior is clear.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Three sentences, each providing essential information: what the tool does, what it returns, and when to use it. No wasted words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no output schema, the description could mention return format details (e.g., ranked list, includes IDs). However, the context of 11 sibling tools and the clear usage guideline make it sufficient for an agent to use correctly. The limit parameter is documented in the schema, so it's fine.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%, so the schema already documents both parameters well. The description adds value by explaining the query parameter's use as 'natural language description' and provides examples like 'analyze housing market trends', which goes beyond the schema's generic description.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description explicitly states 'Search the Pipeworx tool catalog by describing what you need' and 'Returns the most relevant tools with names and descriptions,' clearly identifying the action and resource. It also distinguishes itself from sibling tools by being a meta-tool for discovery, not a domain-specific tool.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides explicit guidance: 'Call this FIRST when you have 500+ tools available and need to find the right ones for your task.' This tells the agent exactly when to use it and implies that other tools are for specific tasks after discovery.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
forgetBInspect
Delete a stored memory by key.
| Name | Required | Description | Default |
|---|---|---|---|
| key | Yes | Memory key to delete |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description must disclose behavior. It correctly indicates the tool deletes data, but does not mention side effects (e.g., permanent deletion, error handling for non-existent keys) or authorization needs. The description is minimal but not contradictory.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single sentence, 6 words, which is concise and front-loaded. However, it lacks additional context that might be useful (e.g., 'Note: this action cannot be undone').
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a simple deletion tool with one parameter and no output schema, the description is adequate but minimal. It does not explain what happens on success/failure, nor does it mention that deletion is irreversible. Given the tool's simplicity, this is acceptable but not exemplary.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100% with one parameter 'key' described as 'Memory key to delete'. The description does not add beyond the schema, but the schema is sufficient. Since coverage is high, baseline 3 is appropriate; a slight bonus is given for clarity of the parameter purpose.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the action ('Delete') and the resource ('a stored memory by key'). It distinguishes the tool from siblings like 'remember' (create) and 'recall' (read), but does not explicitly differentiate from other deletion-like tools, though none appear to exist.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives. For instance, it does not mention that the key must exist before deletion or suggest verifying existence with 'recall' first.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
recallAInspect
Retrieve a previously stored memory by key, or list all stored memories (omit key). Use this to retrieve context you saved earlier in the session or in previous sessions.
| Name | Required | Description | Default |
|---|---|---|---|
| key | No | Memory key to retrieve (omit to list all keys) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description must bear burden. States listing behavior when key omitted, but doesn't detail side effects or persistence details.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences, clear and front-loaded. Every part adds value.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Adequate for a simple retrieval tool with optional parameter. No output schema needed given simple return type implied.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100% and description adds meaning: explains purpose of omitting key (list all) vs specifying it.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Clearly states it retrieves a memory by key or lists all memories. Distinguishes from 'remember' and 'forget' siblings.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Provides some guidance on when to use (retrieve context), but does not explicitly mention when not to use or alternatives.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
rememberAInspect
Store a key-value pair in your session memory. Use this to save intermediate findings, user preferences, or context across tool calls. Authenticated users get persistent memory; anonymous sessions last 24 hours.
| Name | Required | Description | Default |
|---|---|---|---|
| key | Yes | Memory key (e.g., "subject_property", "target_ticker", "user_preference") | |
| value | Yes | Value to store (any text — findings, addresses, preferences, notes) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations, the description carries the full burden. It discloses persistence behavior ('Authenticated users get persistent memory; anonymous sessions last 24 hours'), which is useful. However, it does not mention limits on memory size, key uniqueness, or whether overwriting is allowed, leaving some behavioral gaps.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is three sentences, each adding value: first defines the action, second explains usage, third clarifies persistence. No wasted words, front-loaded with the core purpose.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's simplicity (2 string params, no output schema, no annotations), the description is nearly complete. It covers purpose, usage, and persistence. A slight gap is the lack of mention of overwrite behavior or memory limits, but it's sufficient for most agents.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%, so the baseline is 3. The description adds minimal parameter semantics beyond what the schema provides (e.g., 'key' and 'value' are already described). It does not elaborate on format constraints or usage tips, but the schema is sufficiently clear.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the action ('store'), the resource ('key-value pair in session memory'), and the purpose ('save intermediate findings, user preferences, or context across tool calls'). It distinguishes from siblings like 'forget' and 'recall' by focusing on storage.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description says 'Use this to save intermediate findings, user preferences, or context across tool calls', providing clear usage context. It does not explicitly exclude other tools, but the purpose is specific enough that an agent can infer when to use it (e.g., not for retrieving or deleting).
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
twilio_get_messageBInspect
Get details of a specific Twilio message by its SID.
| Name | Required | Description | Default |
|---|---|---|---|
| sid | Yes | Message SID (starts with SM...) | |
| _authToken | Yes | Twilio Auth Token | |
| _accountSid | Yes | Twilio Account SID |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Annotations are empty, so the description bears full responsibility. It states 'Get details', implying a read-only operation, but does not mention any side effects, rate limits, or required permissions beyond the auth tokens in the schema. Lacks depth but does not contradict any annotations.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, clear sentence that efficiently conveys the tool's purpose without unnecessary words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the simplicity of the tool (3 parameters, no output schema), the description is adequate. It covers the basic purpose, but could mention that the response contains message details (since no output schema exists) or that the SID is required.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%, and each parameter has a description in the schema. The tool description adds no additional parameter information beyond what the schema provides, so a baseline score of 3 is appropriate.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the action ('get details') and the resource ('a specific Twilio message by its SID'). It is specific and informative, though it does not explicitly distinguish from sibling tools like twilio_list_messages.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance on when to use this tool versus alternatives. It does not mention that the user needs the message SID or that it is for retrieving a single message, which could be inferred, but no explicit when-to-use or when-not-to-use advice is given.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
twilio_list_callsBInspect
List recent phone calls from your Twilio account. Returns call SID, status, duration, and direction.
| Name | Required | Description | Default |
|---|---|---|---|
| limit | No | Max calls to return (default 20, max 100) | |
| _authToken | Yes | Twilio Auth Token | |
| _accountSid | Yes | Twilio Account SID |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Annotations are empty, so description carries full burden. It states the tool lists 'recent' calls, implying a time-based default, and mentions returned fields. However, it does not clarify if the tool is read-only, whether authentication parameters are required each call, or any rate limits. The lack of contradiction with annotations (which are empty) means no penalty.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences efficiently convey purpose and return fields. No unnecessary words, but could be slightly more structured (e.g., separate purpose and output).
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a simple list tool with no output schema, the description covers core purpose and output fields. However, it omits details like pagination, default ordering, and whether 'recent' is based on date. Sibling tools and context signals suggest moderate complexity, so a bit more detail would help.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100% with descriptions for all parameters. The description adds no further meaning beyond the schema (e.g., does not explain the 'limit' parameter's effect or how to use account SID/token). Baseline 3 is appropriate since schema already documents parameters.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the action ('List recent phone calls'), resource ('from your Twilio account'), and key return fields ('call SID, status, duration, and direction'). It distinguishes from sibling tools like twilio_make_call and twilio_list_messages by focusing on listing calls.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives. Sibling tools exist for sending SMS or making calls, but no criteria (e.g., need to filter vs. list all) are mentioned.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
twilio_list_messagesAInspect
List recent SMS/MMS messages from your Twilio account. Supports filtering by to/from number and pagination.
| Name | Required | Description | Default |
|---|---|---|---|
| to | No | Filter by destination phone number | |
| from | No | Filter by sender phone number | |
| limit | No | Max messages to return (default 20, max 100) | |
| _authToken | Yes | Twilio Auth Token | |
| _accountSid | Yes | Twilio Account SID |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the burden. It discloses filtering and pagination behavior but does not mention authentication requirements (though account SID and auth token are required params, not explained in description), rate limits, or default/max limits. The default limit of 20 and max 100 are stated in the schema but not in the description.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single sentence, front-loaded with the main purpose. It is concise and clear, though it could benefit from a brief note on required authentication parameters.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool has 5 parameters, no output schema, and no annotations, the description is somewhat complete but lacks details on authentication (though params indicate), response format, and error handling. It is adequate for a simple listing tool but could be more comprehensive.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the schema already documents all parameters. The description adds value by mentioning 'filtering by to/from number' and 'pagination', which maps to the 'to', 'from', and 'limit' parameters, but does not add new meaning beyond the schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the verb 'List' and resource 'SMS/MMS messages from your Twilio account', and distinguishes from sibling tools like twilio_send_sms (sending) and twilio_get_message (single message) by specifying 'list' and mentioning filtering and pagination.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description indicates usage for listing messages with filtering by to/from and pagination, but does not explicitly state when not to use it (e.g., for sending messages) or name alternatives directly. However, the sibling tool names provide context.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
twilio_make_callAInspect
Initiate a phone call via Twilio. Requires a TwiML URL or application SID to control call behavior.
| Name | Required | Description | Default |
|---|---|---|---|
| to | Yes | Destination phone number in E.164 format | |
| url | Yes | TwiML URL that controls what happens when the call connects | |
| from | Yes | Twilio phone number to call from in E.164 format | |
| _authToken | Yes | Twilio Auth Token | |
| _accountSid | Yes | Twilio Account SID |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description carries full burden. It discloses the need for TwiML URL and account/auth parameters, but doesn't mention side effects (cost, phone network delays), rate limits, or error conditions.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences, front-loaded with action and requirement. Efficient and clear.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given 5 parameters, all required, and no output schema, description is adequate but could include info about call status responses or how to retrieve call results. Lacks guidance on async nature.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%, so baseline is 3. Description adds value by explaining that 'url' is a TwiML URL controlling call behavior, and 'from' is a Twilio phone number, which clarifies purpose beyond the schema's basic type descriptions.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description clearly states the action (initiate a phone call), the resource (via Twilio), and adds key requirement (TwiML URL or application SID). Distinct from sibling tools like twilio_send_sms (SMS) and twilio_list_calls (list).
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Description implies it's for making calls, but does not explicitly state when to use this tool versus alternatives like twilio_send_sms. No guidance on prerequisites or alternatives.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
twilio_send_smsBInspect
Send an SMS message via Twilio. Returns the message SID and status.
| Name | Required | Description | Default |
|---|---|---|---|
| to | Yes | Destination phone number in E.164 format (e.g., +15551234567) | |
| body | Yes | Message body text (max 1600 characters) | |
| from | Yes | Twilio phone number to send from in E.164 format | |
| _authToken | Yes | Twilio Auth Token | |
| _accountSid | Yes | Twilio Account SID (starts with AC...) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations, the description partially discloses behavior: it sends an SMS and returns the SID and status. However, it does not mention potential side effects (cost, delivery delays), rate limits, or failure modes (e.g., invalid numbers). Annotations are absent, so the description carries full burden and could be more detailed.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is concise at one sentence plus a return value note. It is front-loaded with the action. However, it could be slightly more informative without being verbose.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool has 5 parameters (all required) and no output schema, the description is somewhat complete but lacks context about authentication requirements, character limits, or error handling. The sibling tools list includes twilio_get_message, which could be compared, but no guidance is given.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%, so the description adds minimal value beyond the schema. It mentions the return value but does not explain parameter details like body character limit or E.164 format beyond what the schema already provides. Baseline 3 is appropriate.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the action ('Send an SMS message'), the resource ('via Twilio'), and the return values ('message SID and status'). It distinguishes from siblings like twilio_make_call and twilio_get_message, but could be more specific about which sibling tools are related.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives like twilio_list_messages or when not to use it (e.g., for sending to international numbers). The description does not mention prerequisites like having a Twilio account with SMS capabilities.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
Claim this connector by publishing a /.well-known/glama.json file on your server's domain with the following structure:
{
"$schema": "https://glama.ai/mcp/schemas/connector.json",
"maintainers": [{ "email": "your-email@example.com" }]
}The email address must match the email associated with your Glama account. Once published, Glama will automatically detect and verify the file within a few minutes.
Control your server's listing on Glama, including description and metadata
Access analytics and receive server usage reports
Get monitoring and health status updates for your server
Feature your server to boost visibility and reach more users
For users:
Full audit trail – every tool call is logged with inputs and outputs for compliance and debugging
Granular tool control – enable or disable individual tools per connector to limit what your AI agents can do
Centralized credential management – store and rotate API keys and OAuth tokens in one place
Change alerts – get notified when a connector changes its schema, adds or removes tools, or updates tool definitions, so nothing breaks silently
For server owners:
Proven adoption – public usage metrics on your listing show real-world traction and build trust with prospective users
Tool-level analytics – see which tools are being used most, helping you prioritize development and documentation
Direct user feedback – users can report issues and suggest improvements through the listing, giving you a channel you would not have otherwise
The connector status is unhealthy when Glama is unable to successfully connect to the server. This can happen for several reasons:
The server is experiencing an outage
The URL of the server is wrong
Credentials required to access the server are missing or invalid
If you are the owner of this MCP connector and would like to make modifications to the listing, including providing test credentials for accessing the server, please contact support@glama.ai.
Discussions
No comments yet. Be the first to start the discussion!