Skip to main content
Glama

list-todo-linked-resources

Read-only

Retrieve linked resources (emails, URLs) for a specific Microsoft To Do task. Each resource includes display name, web URL, application name, and external ID.

Instructions

Get information of one or more items in a partner application, based on which a specified task was created. The information is represented in a linkedResource object for each item. It includes an external ID for the item in the partner application, and if applicable, a deep link to that item in the application.

đź’ˇ TIP: Lists resources linked to a To Do task (emails, URLs, etc.). Each linked resource has displayName, webUrl, applicationName, and externalId.

Input Schema

TableJSON Schema
NameRequiredDescriptionDefault
topNoPage size (Graph $top). Start small (e.g. 5–15) so responses fit the model context; raise only if needed. Use $select to return fewer fields per item. For more rows, use @odata.nextLink from the response instead of a very large $top.
skipNoItems to skip for pagination. Not supported with $search.
searchNoKQL search query — wrap value in double quotes. Cannot combine with $filter.
filterNoOData filter expression. Add $count=true for advanced filters (flag/flagStatus, contains()). Cannot combine with $search.
countNoSet true to enable advanced query mode (ConsistencyLevel: eventual). Required for complex $filter on flag/flagStatus or contains().
orderbyNoSort expression, e.g. receivedDateTime desc
selectNoComma-separated fields to return, e.g. id,subject,from,receivedDateTime
expandNoExpand related entities
todoTaskListIdYesPath parameter: todoTaskListId
todoTaskIdYesPath parameter: todoTaskId
fetchAllPagesNoFollow @odata.nextLink and merge up to 100 pages into one response. Can return enormous payloads—only when the user explicitly needs a full export. Prefer a small $top first, then paginate or narrow with $filter/$search.
includeHeadersNoInclude response headers (including ETag) in the response metadata
excludeResponseNoExclude the full response body and only return success or failure indication
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Annotations indicate readOnlyHint=true and destructiveHint=false, and the description aligns by stating 'Get information'. It adds that each item includes external ID and deep link. No contradictions. However, it does not mention pagination behavior or rate limits beyond schema parameters.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is two short paragraphs plus a tip, each serving a purpose: the first explains the core function, the second provides a useful tip. No wasted words, though the tip could be integrated. Efficient overall.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given 13 parameters, no output schema, and annotations present, the description covers the core purpose and output fields via the tip. However, it lacks details on pagination behavior, result limits, or how to handle large datasets. The schema parameter descriptions fill some gaps, but the description could be more comprehensive.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100%, so baseline is 3. The description adds value by explaining the output structure (displayName, webUrl, applicationName, externalId) beyond parameter descriptions, compensating for lack of output schema. This helps parameter understanding indirectly.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states it retrieves information about linked resources for a specified To Do task, specifying it covers external IDs and deep links. The tip further clarifies it includes displayName, webUrl, etc., distinguishing it from sibling 'create-todo-linked-resource'.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage for retrieving linked resources of a task via the context (e.g., 'based on which a specified task was created'), but does not explicitly state when to use versus alternatives or provide exclusions. A clear mention of when not to use would improve it.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

Install Server

Other Tools

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/Softeria/ms-365-mcp-server'

If you have feedback or need assistance with the MCP directory API, please join our Discord server