Cypress Cloud MCP
Server Details
Direct access to Cypress tests results and accessibility reports in your AI workflow.
- Status
- Healthy
- Last Tested
- Transport
- Streamable HTTP
- URL
- Repository
- cypress-io/ai-toolkit
- GitHub Stars
- 4
Glama MCP Gateway
Connect through Glama MCP Gateway for full control over tool access and complete visibility into every call.
Full call logging
Every tool call is logged with complete inputs and outputs, so you can debug issues and audit what your agents are doing.
Tool access control
Enable or disable individual tools per connector, so you decide what your agents can and cannot do.
Managed credentials
Glama handles OAuth flows, token storage, and automatic rotation, so credentials never expire on your clients.
Usage analytics
See which tools your agents call, how often, and when, so you can understand usage patterns and catch anomalies.
Tool Definition Quality
Average 3.4/5 across 5 of 5 tools scored.
Each tool targets distinct data types (failed tests, flaky tests, projects, runs, feedback) with clear boundaries. The distinction between failed and flaky tests is explicit, and the meta feedback tool is clearly separate from domain operations.
Four tools follow a consistent `cypress_get_*` pattern, but `submit_mcp_feedback` breaks convention entirely by omitting the `cypress_` prefix and using a different verb style, creating inconsistency in the tool set.
Five tools is appropriate for a focused read-only monitoring use case, though at the minimal end of acceptable range. It avoids bloat while covering key query operations.
Severely limited to read-only operations with no ability to trigger runs, cancel executions, or retrieve logs/artifacts (explicitly noted as missing). Lacks full CRUD lifecycle coverage for CI/CD workflows.
Available Tools
5 toolscypress_get_failed_testsGet Failed TestsCInspect
Returns list of failed tests from a given cypress run.
| Name | Required | Description | Default |
|---|---|---|---|
| page | No | Page number to return (default: 1) | |
| limit | No | Number of items to return per page (default: 10, max: 25) | |
| run_url | No | Link to the Cypress Cloud run. --- When used, do not provide `projectId` or `runNumber`. | |
| project_id | No | Cypress Cloud Project Id, usually found in the cypress.config.js file. --- When used, you must also provide `runNumber`. | |
| run_number | No | Cypress Cloud Run number. --- When used, you must also provide `projectId`. |
Output Schema
| Name | Required | Description |
|---|---|---|
| data | No | |
| error | No |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden but offers minimal behavioral context. It fails to mention that results are paginated (despite page/limit parameters), does not describe what constitutes a 'failed' test versus other states, and omits error handling behavior for invalid run identifiers.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Single sentence is front-loaded with the verb and efficiently structured with no filler words. However, given the tool's complexity (5 parameters with mutual exclusivity rules), the description is arguably undersized rather than optimally concise.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Minimum viable given rich structured data: 100% schema coverage and output schema exist, reducing burden on description. However, it lacks explanation of the parameter combination requirements (run_url vs project_id+run_number), which is essential context for successful invocation.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, establishing a baseline of 3. The description adds no parameter-specific guidance, but the schema fully documents parameter relationships (e.g., 'When used, do not provide projectId') and constraints, so no additional compensation is required from the description.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Clear verb (Returns) and resource (failed tests) with scoping (from a given cypress run). However, it does not explicitly differentiate from sibling cypress_get_flaky_tests, which could confuse agents about whether flaky tests are included or excluded from 'failed' results.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Provides no guidance on when to use versus alternatives, nor does it explain the critical mutual exclusivity between run_url and project_id/run_number parameters (which is only documented in individual parameter descriptions). No prerequisites or error conditions mentioned.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
cypress_get_flaky_testsGet Flaky TestsCInspect
Returns list of flaky tests from a given cypress run.
| Name | Required | Description | Default |
|---|---|---|---|
| page | No | Page number to return (default: 1) | |
| limit | No | Number of items to return per page (default: 10, max: 25) | |
| run_url | No | Link to the Cypress Cloud run. --- When used, do not provide `projectId` or `runNumber`. | |
| project_id | No | Cypress Cloud Project Id, usually found in the cypress.config.js file. --- When used, you must also provide `runNumber`. | |
| run_number | No | Cypress Cloud Run number. --- When used, you must also provide `projectId`. |
Output Schema
| Name | Required | Description |
|---|---|---|
| data | No | |
| error | No |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, leaving the description to carry full disclosure burden. It only states the tool 'returns' data, offering no information about what defines 'flaky,' external API calls required, pagination behavior, or permission requirements despite the tool likely accessing Cypress Cloud APIs.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Extremely concise at 8 words with no redundancy. However, the brevity is arguably excessive given the tool's complexity (5 parameters with mutual exclusivity constraints), leaving crucial usage patterns undocumented in the description text.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
The description acknowledges the need to target a 'given cypress run,' but with zero required parameters and complex alternative identifier schemes in the schema, the description should explicitly note the two mutually exclusive identification methods. Output schema exists, reducing burden on description for return values.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
With 100% schema description coverage and detailed parameter descriptions (including conditional logic for run_url vs project_id), the schema carries the semantic load. The description adds no parameter-specific guidance, warranting the baseline score of 3.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly identifies the action (returns/retrieves) and specific resource (flaky tests) with scope (from a given cypress run). It distinguishes the target resource by name, though it doesn't explicitly differentiate from sibling tool 'cypress_get_failed_tests' in the prose.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No explicit guidance on when to use this tool versus siblings (e.g., when to query flaky vs failed tests). No mention of the critical mutual exclusivity between 'run_url' and 'project_id'+'run_number' parameter groups, which is only documented in the schema descriptions.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
cypress_get_projectsGet Cypress ProjectsBInspect
Returns a list of projects available to the user.
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Output Schema
| Name | Required | Description |
|---|---|---|
| data | No | |
| error | No |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description carries full burden. Only behavioral hint is 'available to the user' implying auth scope. Missing details on pagination, caching, or whether this includes archived/deleted projects.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Single sentence with no filler. Front-loaded with the action and resource. Appropriate length for a zero-parameter list operation.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Adequate given low complexity (0 params, flat structure) and presence of output schema. However, gaps remain in explaining workflow relationship to sibling tools that likely require these project identifiers.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Zero parameters per input schema. As per rubric baseline, tools with no parameters score 4 since no semantic clarification is required.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Clear verb ('Returns') and resource ('projects'), with scope ('available to the user'). However, it does not explicitly differentiate from sibling tools like 'cypress_get_runs' that also return lists of entities.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Provides no guidance on when to use this tool versus siblings, nor does it mention that this is likely a prerequisite discovery step needed to obtain project IDs for 'cypress_get_runs' or 'cypress_get_failed_tests'.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
cypress_get_runsGet Cypress RunsAInspect
Retrieves high-level run results, including status, test counts, and tags. It does not return logs, artifacts, or detailed test results.
| Name | Required | Description | Default |
|---|---|---|---|
| page | No | Page number to return (default: 1) | |
| limit | No | Number of items to return per page (default: 10, max: 25) | |
| run_url | No | Link to the Cypress Cloud run. --- When used, do not provide `projectId` or `runNumber`. | |
| git_branch | No | Git branch name to filter runs | |
| git_commit | No | Git commit SHA to filter runs (full or partial, min 7 chars) | |
| project_id | No | Cypress Cloud Project Id, usually found in the cypress.config.js file. --- When used, you must also provide `runNumber`. | |
| run_number | No | Cypress Cloud Run number. --- When used, you must also provide `projectId`. |
Output Schema
| Name | Required | Description |
|---|---|---|
| data | No | |
| error | No |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations, the description carries full burden. It successfully discloses data scope (returns status/counts/tags vs excludes logs/artifacts), but fails to declare operational traits: no mention of read-only safety, pagination behavior (despite page/limit params), or what occurs when called with zero required parameters.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences with zero waste. Front-loaded with the action ('Retrieves...'), followed immediately by scope clarification. The negative constraint in sentence two earns its place by preventing mis-selection.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Adequate for a query tool with rich schema (100% coverage) and output schema. The description covers the essential behavioral contract (high-level summary only). Minor gap: could note pagination behavior given the page/limit parameters, though output schema likely handles return structure.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema has 100% description coverage with clear parameter semantics and cross-parameter constraints already documented (e.g., 'When used, do not provide...'). Description adds no parameter-specific details, which is appropriate given the baseline expectation for high-coverage schemas.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Excellent specificity: 'Retrieves high-level run results' uses a concrete verb with clear resource scope. The second sentence explicitly differentiates from siblings (cypress_get_failed_tests) by stating what is NOT returned (logs, artifacts, detailed results), preventing tool selection errors.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Provides implied usage through negative constraints ('does not return logs...'), which hints when to avoid this tool, but lacks explicit guidance like 'For detailed test logs, use cypress_get_run_details instead' or prerequisite conditions despite having complex parameter dependencies in the schema.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
submit_mcp_feedbackSubmit MCP FeedbackAInspect
Submits user feedback regarding the MCP service.
If the user wishes to do something that is not possible via the tools provided, summarize the request and submit feedback.
| Name | Required | Description | Default |
|---|---|---|---|
| feedback | Yes | User feedback text to submit |
Output Schema
| Name | Required | Description |
|---|---|---|
| data | No | |
| error | No |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description carries full burden. It discloses the trigger condition (capability gaps) but lacks mechanical details: persistence guarantees, idempotency, rate limits, or confirmation behavior. For a mutation operation with no safety annotations, additional behavioral context would be valuable.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences with zero waste. First sentence establishes purpose; second provides usage condition. Information is front-loaded and every sentence earns its place.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given 100% schema coverage, single simple parameter, and existence of output schema (per context signals), the description comprehensively covers necessary invocation context. Appropriate for a straightforward feedback submission tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100% (feedback parameter well-documented). Description adds crucial semantic requirement 'summarize the request', instructing the agent to synthesize user intent rather than relay raw text. This elevates it above baseline schema documentation.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description uses specific verb 'Submits' with clear resource 'user feedback regarding the MCP service'. It effectively distinguishes itself from Cypress data-retrieval siblings (cypress_get_*) by targeting a completely different domain (MCP service meta-feedback vs. Cypress test data).
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Provides explicit when-to-use trigger: 'If the user wishes to do something that is not possible via the tools provided'. This clearly signals the condition for invocation. Missing explicit 'when-not-to-use' or alternative tools, though the condition is sufficiently specific.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
Claim this connector by publishing a /.well-known/glama.json file on your server's domain with the following structure:
{
"$schema": "https://glama.ai/mcp/schemas/connector.json",
"maintainers": [{ "email": "your-email@example.com" }]
}The email address must match the email associated with your Glama account. Once published, Glama will automatically detect and verify the file within a few minutes.
Control your server's listing on Glama, including description and metadata
Access analytics and receive server usage reports
Get monitoring and health status updates for your server
Feature your server to boost visibility and reach more users
For users:
Full audit trail — every tool call is logged with inputs and outputs for compliance and debugging
Granular tool control — enable or disable individual tools per connector to limit what your AI agents can do
Centralized credential management — store and rotate API keys and OAuth tokens in one place
Change alerts — get notified when a connector changes its schema, adds or removes tools, or updates tool definitions, so nothing breaks silently
For server owners:
Proven adoption — public usage metrics on your listing show real-world traction and build trust with prospective users
Tool-level analytics — see which tools are being used most, helping you prioritize development and documentation
Direct user feedback — users can report issues and suggest improvements through the listing, giving you a channel you would not have otherwise
The connector status is unhealthy when Glama is unable to successfully connect to the server. This can happen for several reasons:
The server is experiencing an outage
The URL of the server is wrong
Credentials required to access the server are missing or invalid
If you are the owner of this MCP connector and would like to make modifications to the listing, including providing test credentials for accessing the server, please contact support@glama.ai.
Discussions
No comments yet. Be the first to start the discussion!