Skip to main content
Glama

list-test-runs

Retrieve a list of test runs for a specified project with optional pagination and limit parameters to manage large result sets.

Instructions

List test runs for a project

Input Schema

TableJSON Schema
NameRequiredDescriptionDefault
projectIdYesThe project ID to list test runs for
limitNoMaximum number of results to return (default: 50)
offsetNoNumber of results to skip for pagination (default: 0)
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations provided. Description does not disclose read-only nature or any behavioral traits (e.g., rate limits, data scope). Assumed read-only but not stated.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Very concise (6 words). Front-loads purpose. Could add pagination details without bloat, but acceptable for a simple list tool.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

No output schema; description does not mention pagination behavior, defaults, or return format. Inadequate for a tool with multiple parameters.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100%, so description adds no value beyond schema. Baseline 3 is appropriate; description just restates purpose.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

Clear verb+resource: 'List test runs for a project'. Distinguishes from singular get-test-run and create-test-run. However, it doesn't explicitly mention pagination or scope.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance on when to use vs alternatives (e.g., get-test-run) or context like pagination. Does not specify typical use case.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

Install Server

Other Tools

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/QAStudio-Dev/mcp-server'

If you have feedback or need assistance with the MCP directory API, please join our Discord server