ChiefLab
Server Details
Open execution contract for agents doing business work. 6 operators following one spec. Approval-gated, tenant-isolated, MIT. Reference at chieflab.io/spec/v0.1.
- Status
- Unhealthy
- Last Tested
- Transport
- Streamable HTTP
- URL
Glama MCP Gateway
Connect through Glama MCP Gateway for full control over tool access and complete visibility into every call.
Full call logging
Every tool call is logged with complete inputs and outputs, so you can debug issues and audit what your agents are doing.
Tool access control
Enable or disable individual tools per connector, so you decide what your agents can and cannot do.
Managed credentials
Glama handles OAuth flows, token storage, and automatic rotation, so credentials never expire on your clients.
Usage analytics
See which tools your agents call, how often, and when, so you can understand usage patterns and catch anomalies.
Tool Definition Quality
Average 3.9/5 across 91 of 91 tools scored. Lowest: 2.3/5.
Multiple tools serve overlapping purposes, such as chieflab_post, chiefmo_generate_social_posts, and chiefmo_launch_product. Many tools have aliases (e.g., chieflab_brain_read/chiefmo_company_brain) that create confusion. An agent would struggle to distinguish the correct tool for a given task.
Tool names use a mix of chieflab_ and chiefmo_ prefixes inconsistently. While all are snake_case, the prefixes vary and some verbs are vague (e.g., process, run). However, many names are descriptive and follow a pattern like verb_noun for chieflab tools.
91 tools is extremely high for a single server. Even for a comprehensive marketing platform, this far exceeds typical scopes (3-15 tools). The sheer number makes it difficult for agents to navigate and select the appropriate tool, leading to high cognitive load.
The tool set covers a wide range of marketing operations including launch, social, email, analytics, connectors, and tenant management. Minor gaps exist (e.g., no tool to delete a tenant or update workspace settings), but overall the surface is comprehensive for the intended domain.
Available Tools
91 toolschieflab_brain_readAInspect
Operator: chieflab-brain (primary read). Returns the per-tenant company brain: brand voice samples, repo facts, proof assets, channel performance memory, claim risk log. Drives cross-launch grounding. Spec-renamed alias of chiefmo_company_brain — both names route to the same handler. Record tools: chiefmo_record_proof / chiefmo_record_brand_voice / chiefmo_record_repo_fact / chiefmo_record_channel_result.
| Name | Required | Description | Default |
|---|---|---|---|
| tenantId | No | ||
| workspaceId | No |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description carries full burden. It explicitly states the tool is a 'primary read' and lists the data it returns, assuring the agent it is a read-only operation. While no side effects are described, the purpose as a read tool is clear and consistent.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single paragraph that efficiently communicates purpose, data returned, alias, and related tools. However, it could be better structured for quick scanning (e.g., bullet points for the data list).
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
With no output schema, the description partially compensates by listing what is returned, but lacks detail on structure, pagination, or error conditions. The tool returns multiple complex objects, and more guidance would improve completeness.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 2 parameters (tenantId, workspaceId) with 0% description coverage, and the description provides no guidance on their values or format. This forces the agent to infer usage from context, which is insufficient for correct invocation.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description clearly identifies the tool as the primary read for the per-tenant company brain, enumerating the data categories returned (brand voice samples, repo facts, proof assets, etc.), and distinguishes it from sibling record tools and the aliased chiefmo_company_brain.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Description states it 'drives cross-launch grounding' suggesting when to use, and mentions record tools separately implying they handle writes. However, it does not explicitly contrast with other read tools like chiefmo_check_claims or provide when-not-to-use conditions.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_connect_connectorAInspect
USE WHEN the user wants to connect a third-party data source via OAuth (HubSpot, Stripe, Intercom, Linear, etc.). Returns an authorize URL — surface to user → they grant access → connector becomes live for the workspace and ChiefLab reads real data instead of mock snapshots. NOTE: only chieflab_connect_publish_account (Zernio for social posting) is fully wired in v1; other connectors are scaffolded but the live-data path may be partial.
| Name | Required | Description | Default |
|---|---|---|---|
| provider | Yes | Connector provider id (hubspot, stripe, intercom, linear, etc.). | |
| workspaceId | No | Optional workspace id (defaults to your API key's workspace). |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations, the description must disclose behavior fully. It explains the OAuth flow, return of an authorize URL, and the lifecycle (user grants access → connector becomes live → real data replaces mock). It does not explicitly state whether the tool is read-only or destructive, but the described effect (creating a connector) implies mutation. This is adequate but could be clearer.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise—two sentences. The first sentence immediately states the purpose and use case. The second provides a crucial limitation note. No filler or redundant information.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's simplicity (2 params, no output schema), the description covers the key points: what it does, what it returns, and the effect. The v1 limitation adds important context. However, it does not mention error conditions (e.g., invalid provider) or what happens if OAuth fails. This is acceptable for a straightforward tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%, but the description adds value: it provides example provider values (hubspot, stripe, etc.) and explains that workspaceId defaults to the API key's workspace, information not present in the schema. This helps an agent understand parameter usage beyond the schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool connects third-party data sources via OAuth and returns an authorize URL. It lists examples (HubSpot, Stripe, etc.). However, it does not explicitly distinguish itself from sibling tools like chieflab_connector_setup_steps or chieflab_connect_provider, which could create ambiguity for an agent.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description starts with 'USE WHEN' providing explicit context. It instructs to surface the URL to the user for OAuth grant. The note about v1 limitations ('only chieflab_connect_publish_account is fully wired') informs when not to rely on live data for other connectors. More explicit alternatives or conditions for using siblings could improve this.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_connector_setup_stepsAInspect
Get step-by-step JIT setup instructions for a connector (zernio | resend | ga4 | search_console | github | cms). Returns prerequisites + ordered steps + verification commands. Use when a launch pack reports a connectorBlocker — pipe these steps to the user.
| Name | Required | Description | Default |
|---|---|---|---|
| provider | Yes | zernio | resend | ga4 | search_console | github | cms |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description must cover behavioral traits. It describes the output (prerequisites + ordered steps + verification commands) but does not disclose whether authentication is required, whether it is read-only, or any rate limits. It leaves 'JIT' unexplained. With no annotations, a 3 is appropriate – adequate but not thorough.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences with no superfluous words. The first sentence states purpose and providers; the second provides usage context. Front-loaded and efficient.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a simple tool with one parameter and no output schema, the description covers purpose, usage trigger, and output content. It lacks explanation of 'JIT' and error handling, but overall it is sufficiently complete for an agent to use correctly.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100% for the single parameter 'provider', and the description lists the allowed values in parentheses, echoing the schema. It does not add new meaning beyond what the schema provides. Baseline for high coverage is 3.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states 'Get step-by-step JIT setup instructions for a connector' and lists the allowed providers (zernio, resend, ga4, etc.). It specifies the return content (prerequisites, ordered steps, verification commands). This distinguishes it from sibling tools like chieflab_connect_connector, which likely performs the actual connection.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description explicitly tells when to use the tool: 'Use when a launch pack reports a connectorBlocker — pipe these steps to the user.' This is a specific use case. It does not explicitly say when not to use or mention alternatives, but the guidance is clear and actionable.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_connector_statusAInspect
USE WHEN the user just completed a connector OAuth flow and you want to confirm it succeeded — 'did my connection work?', 'is my Zernio/HubSpot/Stripe connected?'. Returns status: pending | connected | failed for the given connectionId.
| Name | Required | Description | Default |
|---|---|---|---|
| workspaceId | No | ||
| connectionId | Yes |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It discloses the return value format and that it's a read operation for checking status. However, it does not mention error handling, authentication requirements, or what happens if the connectionId is invalid, leaving some behavioral gaps.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single sentence with a clear usage prefix and example. No unnecessary words, front-loaded with key information.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no output schema and few parameters, the description covers the essential purpose and usage. It lacks details on error cases or edge behaviors, but for a simple status check it is largely sufficient.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema has two parameters with 0% description coverage. The description mentions 'for the given connectionId' but does not explain workspaceId at all. Given the low coverage, the description should provide more clarity on workspaceId's role and usage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool checks connector OAuth status, gives example queries, and lists possible return values. It distinguishes itself from siblings like chieflab_connect_connector and chieflab_connector_setup_steps by focusing on post-OAuth confirmation.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description explicitly begins with 'USE WHEN' and provides specific user queries that trigger this tool. It does not mention when not to use or alternatives, but the context is clear enough for an agent to decide.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_connect_providerAInspect
Connect a third-party provider (Zernio, Resend, GA4, Search Console, HubSpot, Stripe, Linear, Notion, Slack) to this workspace. USE WHEN the user wants to wire up publishing, email sending, or analytics readback. For OAuth providers (ga4 / search_console / hubspot) returns an authorizeUrl the agent surfaces to the user. For API-key providers (zernio / resend) returns instructions for the set-key tool. Without this, publish/send/measure tools return 'configure first' errors.
| Name | Required | Description | Default |
|---|---|---|---|
| provider | Yes | Provider id: zernio | resend | ga4 | search_console | hubspot | stripe | linear | notion | slack |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description carries full burden. Discloses that for OAuth providers it returns an authorizeUrl for the user, and for API-key providers it returns instructions for a separate set-key tool. Implies the tool is a prerequisite for other actions. Reasonably transparent, though could mention if it modifies workspace state permanently.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Single paragraph, starts with the main action, then usage guidance, then behavioral specifics. All sentences earn their place. No fluff. Could be slightly more concise by splitting into bullet points, but appropriate for current format.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
No output schema, so description explains the two possible return types. Covers error scenario ('configure first' errors). Lists all provider options. Sufficient for an agent to use this tool correctly. Minor omission: doesn't specify if the tool checks connectivity or just saves configuration.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100% with a description for 'provider' parameter. The description adds value by grouping providers into OAuth vs API-key categories and explaining the behavioral difference, which helps the agent choose the right response path. This goes beyond the simple enum list in the schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Clearly states 'Connect a third-party provider to this workspace' and lists specific providers (Zernio, Resend, GA4, etc.). Distinguishes from siblings by describing the overall purpose, though could explicitly contrast with chieflab_connect_connector. The USEWHEN clause helps clarify scope.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Provides explicit use-case statement ('USE WHEN the user wants to wire up publishing, email sending, or analytics readback'). Categorizes providers by auth type (OAuth vs API-key) and tells agent what to do with the response. Also warns about prerequisite errors for other tools.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_connect_publish_accountAInspect
USE WHEN the user wants to connect a social account (LinkedIn / X / Threads / Instagram / Facebook / Bluesky / TikTok) so chiefmo_publish_approved_post can post to it. Returns a Zernio OAuth URL — surface to user → they click → connect their account → it appears in chieflab_list_publish_accounts. Standard agent flow: call this with platform='linkedin' → get authUrl → surface to user → user connects → call chieflab_list_publish_accounts to confirm + grab accountId → pass accountId to chiefmo_publish_approved_post. NOTE: Zernio requires a profileId (the workspace container). If you don't pass one, the call fails with 'Profile ID required'. Get it from app.zernio.com/settings/profile, or call chieflab_list_publish_accounts first — the accounts response includes profileId on each item.
| Name | Required | Description | Default |
|---|---|---|---|
| platform | Yes | linkedin | x | threads | instagram | facebook | bluesky | tiktok | |
| profileId | No | Zernio profile id. Required by Zernio's /connect endpoint. Get from app.zernio.com/settings/profile or from chieflab_list_publish_accounts response. | |
| redirectUrl | No | Where Zernio should redirect after the user completes OAuth. Default: chieflab.io/runs (approval surface). | |
| workspaceId | No | Optional workspace id (defaults to your API key's workspace). |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations, the description carries full burden. It discloses that the tool returns a Zernio OAuth URL, requires a profileId (with failure consequences), and defaults redirectUrl. It does not cover idempotency or rate limits, but the core behavioral traits are transparent.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is effectively structured with a clear USE WHEN statement upfront. It is concise while covering purpose, flow, and a critical note. Minor minor redundancy (repeated mention of profileId requirement) but overall well-organized.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no annotations and no output schema, the description is fairly complete: it explains the tool's role in the workflow, required parameters, and a common failure case. It does not describe the exact structure of the returned authUrl (e.g., protocol), but that is implied.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%, and the description adds meaningful context: it explains the role of each parameter in the OAuth flow, especially the critical profileId and the default redirectUrl. This goes beyond the schema descriptions.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description explicitly states the tool's purpose: connecting a social account for posting via chiefmo_publish_approved_post. It lists supported platforms and clearly distinguishes from sibling tools like chieflab_list_publish_accounts by focusing on the connection step.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides a standard agent flow (call this with platform, get authUrl, surface to user, then confirm with list_publish_accounts and pass accountId to publish). It also advises on obtaining profileId, which is critical. This is explicit guidance on when and how to use the tool.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_create_tenantAInspect
USE WHEN you (an AI product builder) are managing multiple end-users and want isolated brand context per user — 'create a tenant for my new customer', 'set up Acme Co as a tenant', 'multi-tenant brand isolation'. Each tenant gets its own brand context, voice samples, memory, scoped under your workspace.
| Name | Required | Description | Default |
|---|---|---|---|
| name | Yes | Display name (e.g. 'Acme Co'). | |
| domain | No | Optional URL/domain (used for auto-context extraction). | |
| industry | No | Optional industry (e.g. 'B2B SaaS'). | |
| tenantId | Yes | Short slug used in MCP calls (e.g. 'tenant_abc123' or 'acme-co'). |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description must cover behavioral traits. It states that each tenant gets its own brand context, voice samples, memory, and is scoped under the workspace. However, it does not mention idempotency, error handling, or what happens on duplicate tenantId.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences, each sentence serves a purpose. The first provides the usage signal, the second explains the outcome. No unnecessary words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
The description explains the purpose and outcome but does not mention the return value (e.g., whether the created tenant object is returned), error conditions, or prerequisite operations. Given no output schema, more details on the result would improve completeness.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the baseline is 3. The description adds context that tenantId is a 'short slug used in MCP calls' and that domain is for 'auto-context extraction', which adds slight value beyond the schema descriptions.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description clearly states the verb 'create' and resource 'tenant', and differentiates from sibling tools like chieflab_list_tenants and chieflab_set_tenant_context by specifying it's for creating a new tenant for isolated brand context.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicitly starts with 'USE WHEN' and describes scenarios like managing multiple end-users and setting up a tenant for a new customer. Does not explicitly mention when not to use or alternatives, but context is clear.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_customer_success_queueAInspect
Internal CS-ops queue: lists blocked launches, missing-approval items, recent connector failures, and due followups across the workspace. Use to triage what's stuck. Filterable by bucket.
| Name | Required | Description | Default |
|---|---|---|---|
| limit | No | ||
| bucket | No | blocked_launches | missing_approvals | connector_failures | due_followups | all (default) | |
| workspaceId | No |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Describes the tool as listing items, implying read-only behavior, but no explicit mention of non-destructive nature or output format. With no annotations, description carries burden; it's adequate but minimal.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two concise sentences front-load the purpose and usage. No unnecessary words, every sentence adds value.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Lacks details on output format (e.g., fields returned) and does not explain limit or workspaceId parameters. For a list tool, more context is needed for proper invocation.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Only bucket parameter has schema description (enum values), and the description merely repeats 'Filterable by bucket'. Limit and workspaceId are not explained beyond schema type, adding no value.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Clearly states it lists blocked launches, missing-approval items, connector failures, and due followups for triage. Distinct from sibling tools which focus on other domains.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicitly says 'Use to triage what's stuck' and mentions filterability by bucket, providing clear usage context. No exclusion or alternative guidance needed as no sibling overlaps.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_get_connect_actionAInspect
P13 — for a single (typically blocked) channel, return the exact next tool to call to unblock it: chieflab_connect_provider for OAuth/key providers, chieflab_use_manual_fallback for manual-only channels. USE WHEN the user clicks an unblock button or the agent decides which channel to wire next.
| Name | Required | Description | Default |
|---|---|---|---|
| channel | Yes | Channel id from the catalog (publish_x / send_email / post_product_hunt / etc.) or a legacy alias (linkedin / email / product_hunt). |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It only states 'return the exact next tool' without disclosing any behavioral traits like side effects, permissions, or return format. This is insufficient for a tool that likely influences subsequent actions.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences, each carrying essential information. The first sentence states core functionality, the second provides usage context. No wasted words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
No output schema exists, so the description should explain return values. It says 'return the exact next tool' but does not specify the return format (e.g., tool name as string or object). It also omits error handling or edge cases. For a simple tool, it's mostly adequate but missing some details.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The single parameter 'channel' has a schema description, and the tool description adds clarification ('Channel id from the catalog... or a legacy alias'), which adds value beyond the schema. Schema coverage is 100%, making this a strong addition.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: returning the next tool to call for an unblock action, and lists two specific sibling tools (chieflab_connect_provider and chieflab_use_manual_fallback). This distinguishes it well from other tools.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description explicitly says 'USE WHEN the user clicks an unblock button or the agent decides which channel to wire next,' providing clear usage context. It does not mention when not to use or alternatives, but the context is sufficient for this decision.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_get_users_after_buildAInspect
Outcome-named alias for chiefmo_launch_product. USE WHEN a coding agent just built, shipped, or deployed something and the user asks 'get users', 'find customers', 'make money from this', 'get paying users', 'get eyes on this', 'what now?', or 'how do I launch this?' This is the agent-dependency tool: after build, call this to turn the product into a repo-grounded money/distribution loop with launchPack, signed reviewUrl, approval-gated publish/send actions, 24h measurement, memory, next move, moneyLoop, and agentDependency. Gather repoContext first. Route exactly like chiefmo_launch_product; do not use chiefmo_diagnose_marketing for a new product.
| Name | Required | Description | Default |
|---|---|---|---|
| goal | No | Desired outcome, e.g. 'Get first 100 users', 'find paying customers', 'get design partners', 'make money from this'. | |
| brand | No | Optional inline brand context. | |
| channels | No | Optional launch channels. Default: LinkedIn, X, Product Hunt, email, landing hero. | |
| tenantId | No | Optional tenant/end-user id for stored brand context and memory. | |
| outputMode | No | context | draft | full. Defaults to context/draft behavior from chiefmo_launch_product. | |
| productUrl | No | Live product URL. Pass either productUrl or productDescription. | |
| repoContext | No | Strongly recommended. Include whatChanged, recentCommits, changedFiles, routes, readme, targetCustomer, launchGoal. | |
| idempotencyKey | No | Optional dedupe key to prevent duplicate launch preparation. | |
| productDescription | No | Short product description if no public URL exists yet. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Although no annotations are provided, the description details the tool's behavior: it sets up a launchPack with approval-gated actions, measurement, memory, etc. It also mentions prerequisites ('Gather repoContext first'). The description effectively communicates the tool's role as an 'agent-dependency tool' without contradicting any annotations.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is front-loaded with the core purpose and structured logically: alias declaration, use cases, what the tool does, prerequisites, and routing. While somewhat verbose, every sentence adds value and is necessary for agent understanding. The length is justified by the tool's complexity.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the absence of output schema and annotations, the description provides a comprehensive overview including the tool's role, output artifacts (launchPack, reviewUrl), lifecycle details (24h measurement, next move), and explicit usage context. It leaves no ambiguity about when and how to invoke the tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 100% description coverage, so the baseline is 3. The description adds minimal additional parameter guidance beyond the schema, except for clarifying the productUrl/productDescription alternative and emphasizing repoContext. The schema itself is already fairly descriptive.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states it is an 'Outcome-named alias for chiefmo_launch_product' and specifies the exact scenarios (after build, when user asks for users/customers/money). It distinguishes itself from siblings by explicitly noting it should not be confused with chiefmo_diagnose_marketing.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides explicit when-to-use criteria: 'USE WHEN a coding agent just built, shipped, or deployed something and the user asks...' along with specific user queries. It also gives routing instructions and a clear alternative ('do not use chiefmo_diagnose_marketing for a new product').
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_launch_productCInspect
Operator: chieflab-launch (primary). Spec v0.1 reference implementation of the orchestrator. End-to-end launch loop: repo-grounded positioning + per-channel drafts (linkedin, x, hn, reddit, email, landing_hero, product_hunt) + image briefs + approval-gated publishActions + signed reviewUrl + 24h measurement queue. Internally composes chieflab-post + chieflab-email under one runId. Spec-renamed alias of chiefmo_launch_product (backwards-compat retained).
| Name | Required | Description | Default |
|---|---|---|---|
| goal | No | ||
| brand | No | ||
| channels | No | ||
| tenantId | No | ||
| productUrl | No | ||
| webhookUrl | No | ||
| repoContext | No | ||
| workspaceId | No | ||
| imagesNeeded | No | ||
| idempotencyKey | No |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It discloses internal composition (post, email under one runId), approval-gating, review URL generation, and 24h measurement queue. However, it does not mention authentication needs, rate limits, or destructive nature beyond implied mutations.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is single-paragraph but front-loads the operator context. However, it is dense and uses spec-revision language ('Spec v0.1', 'spec-renamed alias') that may be verbose without aiding utility. Every sentence contributes but structure could be cleaner.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (10 parameters, nested objects, no output schema, no annotations), the description explains the orchestration flow but lacks specifics on input requirements, error handling, prerequisites, or return value format. Leaves gaps for an AI agent to select and invoke correctly.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 0% despite 10 parameters (some nested). The description hints at channels (linkedin, x, etc.) and image briefs, but does not explicitly map to schema properties like goal, brand, productUrl, or idempotencyKey. High complexity with minimal compensation.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly identifies the tool as an end-to-end launch orchestrator with a specific verb and resource ('launch product'). It mentions composed sub-tools (chieflab-post, chieflab-email) and distinguishes from them, but the dense spec-wording may reduce immediate clarity.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description states it is the 'primary' operator and a spec reference, but does not explicitly say when to use this tool vs. alternatives like chiefmo_launch_product or other launch-related siblings. It notes backwards compatibility with chiefmo_launch_product, leaving confusion about which to invoke.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_list_available_channelsAInspect
P13 — list every channel ChiefLab can ship to (LinkedIn / X / Threads / Bluesky / Instagram / Facebook / TikTok / Product Hunt / Hacker News / Reddit / Discord / Slack / Indie Hackers / DEV.to / YC Bookface / email / blog / GA4 / Search Console / Stripe). Returns the catalog grouped by category (social / community / email / blog / analytics) with each channel's capabilities (publish / schedule / read_metrics / connect / fallback) and providers. USE WHEN the user asks 'what channels can ChiefLab post to?', 'which platforms are supported?', or before calling chiefmo_launch_product to pick the right channels.
| Name | Required | Description | Default |
|---|---|---|---|
| category | No | Optional. Filter to one category: social | community | email | blog | analytics. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations, the description fully discloses that the tool lists channels and returns a grouped catalog with capabilities and providers. It implies a read-only operation and adds value by outlining the output structure. Minor omission: does not mention if data is real-time or cached, but overall transparent.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is concise but packs substantial information: the list of channels, output structure, and usage guidance. The 'P13 —' prefix slightly hinders clarity but is a minor flaw. Overall, it is front-loaded and efficient.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no output schema, the description adequately explains the return format (grouped by category with capabilities and providers). The single parameter is well-documented. The tool's purpose is clear in the context of siblings, making the description complete for agent selection and invocation.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 100% coverage for the single optional 'category' parameter, which is described as filtering by category. The description reinforces this by mentioning grouping by category. The schema already provides sufficient meaning, so the description adds minimal extra value.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool lists every channel ChiefLab can ship to, enumerates specific platforms, and explains the output grouping by category with capabilities. It distinguishes itself from siblings like 'chieflab_list_channel_readiness' by focusing on the full catalog.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description explicitly tells when to use the tool: 'USE WHEN the user asks what channels can ChiefLab post to? or before calling chiefmo_launch_product to pick the right channels.' This provides clear context and an alternative action.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_list_channel_readinessAInspect
P13 — for every channel in the catalog, return its readiness state for THIS workspace: { ready, provider, blocker, fix, manualFallback, fallbackKey }. Drives the dashboard connector-readiness panel and the launch-time blockedActions list. USE WHEN the user asks 'which channels can I publish to right now?' or before chiefmo_launch_product to pre-warn about blocked channels.
| Name | Required | Description | Default |
|---|---|---|---|
| workspaceId | No | Optional workspace id. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, and description does not disclose behavioral traits like side effects, idempotency, or authorization needs. Only states what it returns, lacking transparency beyond that.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Description is front-loaded with purpose and includes useful usage context. Slightly verbose with internal label 'P13' but overall efficient.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no output schema, description lists possible states and applications. Lacks mention of prerequisites or authentication, but sufficient for a read-only list tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema has 100% coverage on one optional parameter. Description adds context by referring to 'THIS workspace' but does not significantly extend schema meaning.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states that the tool returns readiness states for every channel in a catalog for the current workspace, listing possible states and specific use cases. It distinguishes itself from siblings like 'chieflab_list_available_channels'.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicitly tells when to use (e.g., user asking about publishable channels, before chiefmo_launch_product). Does not mention when not to use, but provides clear context.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_list_connectorsAInspect
USE WHEN you want to know which third-party connectors are connected for this workspace — 'what data sources do I have wired?', 'is X connector live?'. Returns all connectors across all operators with OAuth-live status (live data vs. mock snapshots).
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description bears full responsibility. Describes output as 'all connectors across all operators with OAuth-live status,' implying read-only behavior. Lacks details on authorization requirements, pagination, or any side effects.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences, front-loaded with usage scenario, then output description. No wasted words, highly efficient.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a simple list tool with no parameters, description covers purpose and key output detail (OAuth status). Could mention returned fields or format, but not critical. Adequate given sibling tools cover specific actions.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema has zero parameters, so description does not need to add parameter information. Schema coverage is 100% trivial. Baseline for no parameters is 4.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description clearly states the tool lists third-party connectors for the workspace with OAuth-live status. It distinguishes from siblings like 'chieflab_connector_status' by focusing on listing all connectors rather than a single one.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicit 'USE WHEN' phrase provides clear context for which scenarios to use the tool, including example queries. Does not explicitly mention when not to use, but the context is sufficient given the sibling tools.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_list_email_sendersAInspect
USE WHEN you need to know which from addresses chiefmo_send_approved_email will accept — 'list my verified email domains', 'which sender can I use?'. Returns verified Resend sender domains for the workspace. Requires RESEND_API_KEY (env-default or set via chieflab_set_resend_key).
| Name | Required | Description | Default |
|---|---|---|---|
| workspaceId | No | Optional workspace id. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description carries full burden. It states the tool returns verified sender domains, indicating a read-only operation. Mentions API key requirement but lacks details on error handling or quota limits. Adds moderate behavioral context beyond the schema.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences, front-loaded with use case and prerequisite. No redundancy or unnecessary words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given low complexity (one optional param, no output schema), description adequately covers purpose, usage trigger, and key prerequisite. Lacks details on return format or failure modes, but sufficient for a simple list tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100% with one optional workspaceId parameter. Description does not add meaning beyond schema; it only references 'the workspace' which is already in schema. Baseline 3 applies.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description specifies the tool lists verified Resend sender domains for the workspace, with explicit use cases like 'list my verified email domains' and 'which sender can I use?'. This distinguishes it from sibling tools focused on sending emails or managing keys.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Starts with 'USE WHEN' and provides example queries. Also notes prerequisite: requires RESEND_API_KEY either from env or set via another tool. Gives clear context for invocation.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_list_launch_templatesAInspect
List launch templates saved for this workspace (most recently created first). Pass label of one to chiefmo_launch_product as templateId for a one-call repeat launch.
| Name | Required | Description | Default |
|---|---|---|---|
| limit | No | ||
| workspaceId | No |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description carries full burden. It discloses sorting order and relationship to another tool, but does not mention read-only nature, authentication needs, or error conditions.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences, front-loaded with purpose and sorting order. Efficient and to the point without unnecessary words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no output schema and 0% schema coverage, the description is incomplete. It does not explain what the response contains (e.g., template objects), nor details on parameters. It covers sorting and usage link but lacks full context.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 0%, so description should explain parameters. It does not describe 'limit' or 'workspaceId' explicitly, only implies workspace scope. Missing parameter details.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool lists launch templates for the workspace, sorted by most recent first. It also distinguishes from siblings by referencing chiefmo_launch_product, giving specific purpose.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
It explains when to use the tool (to list templates) and how to use the output with another tool. However, it does not explicitly state when not to use it or list alternatives.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_list_playbooksAInspect
List the ChiefLab launch playbooks (devtool / saas / ai_agent / marketplace / internal_tool / consumer) — what's the default channel mix, default angle, ICP, proof expectations, risks-to-avoid for each. Use to show the user options before calling chiefmo_launch_product, or to pick productType explicitly when auto-detection is wrong.
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided; description carries full burden. The verb 'list' implies a read-only, non-destructive operation. While it doesn't explicitly state no side effects, the context of listing playbooks is clearly a safe query. Adequately transparent.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences: first explains what the tool does, second explains usage context. No wasted words, front-loaded with key information.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a simple list tool with no parameters and no output schema, the description fully covers what it returns (playbook details) and when to use it. No missing context.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema has zero parameters, so no parameter description needed. Baseline of 4 is appropriate since the description does not need to add meaning beyond schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description clearly states the tool lists launch playbooks with specific fields (channel mix, angle, ICP, etc.) and explicitly differentiates its use from sibling tools like chiefmo_launch_product. It gives a precise verb+resource description.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Provides explicit when-to-use context: 'before calling chiefmo_launch_product' and when auto-detection is wrong. Implicitly guides to use chiefmo_launch_product after. Could be slightly more explicit about when not to use, but overall clear.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_list_publish_accountsAInspect
USE WHEN you need accountIds before publishing — 'which social accounts are connected?', 'list my LinkedIn/X/Instagram accounts'. Returns connected Zernio accounts (LinkedIn, X, Threads, Instagram, Facebook, etc.) with the accountIds you'll pass to chiefmo_publish_approved_post. Requires ZERNIO_API_KEY in env or via chieflab_set_zernio_key.
| Name | Required | Description | Default |
|---|---|---|---|
| workspaceId | No | Optional. Defaults to the workspace derived from your API key. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description must carry the burden. It discloses that the tool returns connected accounts with accountIds, implying a read-only operation with no side effects. It does not, however, mention error behavior (e.g., missing API key), which would improve transparency.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is remarkably concise: two sentences. The first sentence front-loads the purpose and usage context. Every word adds value; there is no fluff or repetition.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
The tool has no output schema, so the description should explain the return value fully. It states the tool returns 'connected Zernio accounts ... with the accountIds', which is adequate for a simple list. However, it does not specify the structure (e.g., list of objects with platform type, account name) or whether pagination exists. For a tool with one optional parameter, the description is minimally complete.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema has 1 optional parameter (workspaceId) with full description in the schema. Schema description coverage is 100%, so the description adds no extra meaning beyond restating that it's optional and defaults to the workspace from the API key. Baseline 3 is appropriate.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: retrieving accountIds for publishing. It specifies the use case ('USE WHEN you need accountIds before publishing'), lists the social platforms (LinkedIn, X, Threads, etc.), and explains the output's downstream usage (passed to chiefmo_publish_approved_post). This distinguishes it from sibling tools that handle publishing, approval, or other actions.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides explicit usage context with 'USE WHEN' and example queries. It mentions a prerequisite (ZERNIO_API_KEY via environment or previous tool call). However, it does not state when not to use this tool or offer alternatives among siblings, so it loses one point.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_list_tenantsAInspect
USE WHEN you want to see which end-users (tenants) are configured in this workspace — 'list my tenants', 'show me all customers I've onboarded'. Returns active tenants by default; pass status='paused'|'archived' for others.
| Name | Required | Description | Default |
|---|---|---|---|
| status | No | active | paused | archived (default: active) |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden. It discloses the default behavior (active tenants) and filtering options for paused/archived. However, it doesn't mention output format, pagination, or side effects, which is adequate for a simple read operation but not comprehensive.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two well-structured sentences, each sentence earning its place. Clearly front-loaded with purpose and usage examples, no unnecessary words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the simplicity of the tool (single optional parameter, no output schema, no nested objects), the description covers the essential behavior. It lacks details like pagination or return structure, but for a basic list tool, this is complete enough.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100% (one parameter described). The description merely rephrases the schema's 'active | paused | archived (default: active)' as 'pass status='paused'|'archived' for others,' adding minimal new information. Baseline 3 is appropriate.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description clearly states the tool's purpose: listing end-users/tenants configured in the workspace, with example natural language queries. It distinguishes itself from sibling tools like chieflab_create_tenant and chieflab_set_tenant_context by focusing on reading existing tenants.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicitly specifies when to use ('USE WHEN you want to see...') and how to use (default active, pass status for others). Lacks explicit 'when not to use' or alternatives, but the context is clear and sufficient for basic usage.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_measureBInspect
Operator: chieflab-measure (primary). 24h post-launch readback: pulls Zernio engagement, GA4 traffic, Search Console queries; returns structured metrics + next-iteration brief. Spec-renamed alias of chiefmo_post_launch_review — both names route to the same handler.
| Name | Required | Description | Default |
|---|---|---|---|
| runId | Yes | ||
| workspaceId | No |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description must fully disclose behavioral traits. It mentions 'readback' but does not explicitly confirm read-only nature, state authentication requirements, rate limits, or side effects. The description lacks sufficient behavioral transparency beyond the high-level operation.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise, with two short sentences that front-load the core purpose and alias information. No unnecessary words or repetition.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given two parameters, no output schema, and no annotations, the description is incomplete. It does not explain input requirements, output structure, or any constraints. The tool's complexity (multiple data sources, structured metrics, brief) is not fully addressed.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 0%, and the description does not explain any of the parameters (runId, workspaceId). It only describes the output data sources and format without clarifying what the inputs represent or how they affect execution.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description clearly states the tool's purpose as a 24h post-launch readback that pulls specific data sources (Zernio engagement, GA4 traffic, Search Console queries) and returns structured metrics plus a next-iteration brief. It also identifies itself as a spec-renamed alias, providing full clarity.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage context (24h post-launch) but does not explicitly state when to use or not use this tool versus alternatives. There is no guidance on when to prefer this over chiefmo_measure_launch_results or other sibling tools.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_postAInspect
Operator: chieflab-post (primary). Single-channel publish loop. Accepts channel + product context, returns one draft + one publishAction + reviewUrl. Use when the user says "post this to linkedin" / "draft a hacker news submission" / "send a tweet about this" — anywhere the full launch loop is overkill. Approval-gated, spec v0.1 conformant. Falls through to the launch handler with channels=[channel], imagesNeeded=0.
| Name | Required | Description | Default |
|---|---|---|---|
| goal | No | ||
| brand | No | ||
| channel | Yes | linkedin | x | hacker_news | reddit | product_hunt | threads | instagram | |
| tenantId | No | ||
| productUrl | No | ||
| repoContext | No | ||
| workspaceId | No | ||
| idempotencyKey | No |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations, the description discloses that it is approval-gated, returns specific outputs, and falls through to the launch handler. It could mention failure modes or rate limits, but overall it is informative.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is concise and well-structured: operator, function, inputs/outputs, usage, behavioral notes. Every sentence adds value with no redundancy.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Provides a good overview but lacks details on all 8 parameters and precise return types. Given no output schema or annotations, more parameter documentation would improve completeness.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is only 13% (only channel has description). The description vaguely mentions 'product context' but does not detail other parameters like goal, brand, or tenantId. It fails to compensate for the low coverage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states it's a single-channel publish loop that returns a draft, publishAction, and reviewUrl. It distinguishes itself from the 'full launch loop' but does not explicitly differentiate from sibling tools like chiefmo_publish_approved_post.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Provides explicit usage scenarios ('when user says post this to linkedin...') and notes when the full launch loop is overkill. Lacks a when-not-to-use condition but gives clear context.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_record_manual_publishAInspect
USE WHEN the user has manually posted to a channel returned by chieflab_use_manual_fallback (Product Hunt / HN / Reddit / Discord / etc.) and wants to feed the live URL back to ChiefLab so the closed loop continues. Records the URL on the original publishAction (status flips from 'approved' to 'executed_manually'), persists a proof_asset to the P9 company brain, and queues 24-hour metrics readback via chiefmo_post_launch_review. Without this tool, manually-posted channels are lost to ChiefLab's measurement loop.
| Name | Required | Description | Default |
|---|---|---|---|
| notes | No | Optional. Any context on how the post performed in the first hour (early upvotes, top comment, reactions). Persists into the P9 brain. | |
| actionId | Yes | The publishAction.id that the user manually published (from chieflab_use_manual_fallback). Required. | |
| publishedAt | No | Optional ISO 8601 timestamp. Defaults to now. | |
| workspaceId | No | Optional workspace id. | |
| publishedUrl | Yes | The live URL of the manual post (e.g. producthunt.com/posts/<slug>, news.ycombinator.com/item?id=<id>, reddit.com/r/SaaS/comments/<id>). |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Despite no annotations, the description fully discloses behavioral traits: it flips the publishAction status, persists a proof_asset, and queues metrics readback. This gives the agent a clear understanding of side effects and consequences of not using the tool.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, well-structured paragraph. It front-loads the usage condition and explains functionality and consequences concisely. No unnecessary text, but could be broken into multiple sentences for clarity.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the 5-parameter schema and no output schema, the description provides sufficient context: usage scenario, side effects, and importance. It does not describe return values, but the tool's purpose implies it completes an action, and the side effects are well explained.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema has 100% coverage with good descriptions. The tool description adds overall context but does not significantly enhance individual parameter meaning beyond the schema. However, it references the fallback tool for actionId and publishedUrl, slightly elevating utility.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: to record the live URL of a manually posted channel back to ChiefLab. It specifies the verb 'record' and the resource 'manual publish', and distinguishes from siblings by tying to chieflab_use_manual_fallback and explaining the closed loop continuation.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description starts with 'USE WHEN', explicitly stating the condition (user manually posted using chieflab_use_manual_fallback) and the motivation (feed back URL for closed loop). It also warns that without this tool, manually-posted channels are lost, providing clear guidance on when to use.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_record_rendered_copyAInspect
USE WHEN your agent's LLM has rendered the final copy from a launch-pack brief and you want the human reviewer to see the rendered text on the reviewUrl (not the brief). Closes the orchestration loop: ChiefLab returns briefs (cheap), your LLM renders (high-quality), this tool writes the rendered copy back so reviewers approve real text, not prompts. Pass the assetId from launchPack.channels[channel].assetId (or any other ChiefLab asset). Idempotent — second call replaces the prior rendered body.
| Name | Required | Description | Default |
|---|---|---|---|
| body | Yes | The final rendered copy your LLM produced. Markdown is fine. Replaces the brief on the reviewUrl. | |
| assetId | Yes | The asset to attach rendered copy to. Get from launchPack.channels[channel].assetId or run.outputs[i].assetId. | |
| renderedBy | No | Optional. Which model produced this (e.g. 'sonnet-4.6', 'gpt-4o', 'gemini-2.5-pro'). Stored in metadata for audit. | |
| workspaceId | No | Optional workspace id. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations, description fully discloses behavior: it writes to reviewUrl, closes the loop, is idempotent ('second call replaces prior rendered body'). Does not mention potential side effects like overwriting, but idempotency implies safe replacement.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Three sentences, no fluff. Key directive first. Every sentence adds value. Well-structured and front-loaded.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no output schema, description adequately explains purpose, usage, and idempotency. Does not detail return values, but tool is simple write operation. Sufficient for selection and invocation.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%, but description adds meaningful context: body is 'final rendered copy', 'Markdown is fine', 'replaces brief on reviewUrl'. assetId has origin hint. renderedBy explained. Adds value beyond schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description starts with a clear 'USE WHEN' statement specifying the exact scenario. It distinguishes the tool from siblings by explaining the orchestration loop (ChiefLab returns briefs, LLM renders, this tool writes). It uses specific verbs: 'rendered copy', 'writes', 'reviewers approve real text'.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Provides explicit context on when to use ('when your agent's LLM has rendered the final copy'). Describes the orchestration flow and where to get the assetId. Lacks explicit 'when not to use' or alternatives, but the context is clear enough.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_record_voice_sampleAInspect
USE WHEN the user approves or rejects a draft and you want the next run to learn — 'remember this is the right tone', 'never write like this again', 'add this email as a reference'. Stores approved/rejected examples per tenant; future operator prompts ground against approved tone and avoid rejected patterns. The 'rejection becomes memory' loop.
| Name | Required | Description | Default |
|---|---|---|---|
| kind | Yes | approved | rejected | reference | |
| channel | No | email | social | landing | ad | blog (optional) | |
| content | Yes | The actual text | |
| feedback | No | Optional rejection feedback like 'too corporate, make it more founder-led' | |
| tenantId | Yes |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It explains that the tool stores examples per tenant and creates a 'rejection becomes memory loop', implying persistent state changes. However, it lacks details on side effects like overwriting behavior, data limits, or authentication requirements.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is concise and front-loaded with the key use case, though the last sentence ('The 'rejection becomes memory' loop.') adds a metaphorical flourish that could be considered non-essential.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a tool with 5 parameters, no output schema, and no annotations, the description covers purpose and usage but lacks details on how the stored data is later accessed, size constraints, or error handling. It is adequate but not comprehensive.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 80%, and the description adds value by explaining the 'feedback' parameter with an example ('too corporate, make it more founder-led'). However, it does not meaningfully expand beyond the schema's own descriptions for 'kind', 'content', or 'channel'.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly identifies the tool's purpose: storing approved/rejected examples for learning. It uses specific verb+resource ('stores approved/rejected examples per tenant') and distinguishes from siblings by focusing on voice learning, not channel preference or brand voice recording.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description starts with 'USE WHEN the user approves or rejects a draft and you want the next run to learn', providing explicit usage context. However, it does not mention when not to use it or suggest alternatives among the many sibling tools.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_remember_preferred_channelAInspect
P13 — record a workspace's per-job preference (e.g. 'this workspace prefers LinkedIn over X for B2B launches', 'always use the founder LinkedIn account, not company page'). Future runs of chiefmo_launch_product read these preferences before picking the channel mix. Persists into the P9 brain.
| Name | Required | Description | Default |
|---|---|---|---|
| channel | Yes | Channel id (publish_linkedin, send_email, etc.). | |
| preference | Yes | Free-form preference object (e.g. { accountId, persona, weight }). | |
| workspaceId | No |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description bears full responsibility for behavioral disclosure. It states that data persists into the 'P9 brain', implying a write and storage operation. However, it does not specify whether preferences are overwritten or appended, or how workspaceId (an optional parameter) affects behavior. This is adequate but not rich.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is very concise: two sentences front-load the purpose and example, with no extraneous information. Every sentence contributes meaning.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's specific role in recording preferences for chieflab_launch_product, the description is largely complete. It explains the input, persistence, and downstream consumer. However, it lacks details about default behavior when workspaceId is omitted or whether existing preferences are replaced. Still, it is sufficient for an agent to understand when and how to use the tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 67% (two of three parameters have descriptions, but workspaceId's is empty). The tool description provides an example for the 'preference' parameter, adding some value beyond the schema's 'Free-form preference object' label. However, it does not clarify the 'channel' parameter's format or the role of 'workspaceId'. Overall, it compensates moderately for the schema gap.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states that the tool records a per-job preference for a workspace, with concrete examples like 'prefers LinkedIn over X for B2B launches'. The verb 'record' and the specified resource (workspace per-job preference) are unambiguous. This differentiates it from sibling tools like chieflab_launch_product, which reads these preferences.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description explains that future runs of chiefmo_launch_product read these preferences, indicating the tool should be used before launching. It does not explicitly list when not to use it or provide alternatives, but the context of persisting preferences for a specific workflow is clear.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_retry_blocked_actionAInspect
P13 — re-check a blocked publishAction's connector readiness. Returns whether the channel is now ready (so the agent can re-fire chiefmo_publish_approved_post / chiefmo_send_approved_email) or still blocked with the up-to-date connect path. USE WHEN the user has just connected a missing provider and the agent needs to know if the action can fire.
| Name | Required | Description | Default |
|---|---|---|---|
| actionId | Yes | The blocked publishAction.id from chiefmo_launch_product. | |
| workspaceId | No |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description carries full burden. It states the return value (ready or still blocked), but does not disclose side effects (e.g., whether any state is modified) or error conditions. Adequate but not comprehensive.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two concise sentences with no filler. Front-loaded with action and purpose, followed by usage guidance. Every sentence adds value.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no output schema, the description provides return value context. Covers the main use case, but lacks details on error handling or exact response format. Still sufficient for the tool's simplicity.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema describes actionId but not workspaceId (50% coverage). The tool description adds no additional meaning for parameters; workspaceId remains unexplained. The description does not compensate for the schema gap.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Clearly states the verb 're-check' and the resource 'blocked publishAction's connector readiness'. Explicitly distinguishes from sibling tools like chiefmo_publish_approved_post and chiefmo_send_approved_email by mentioning they can be re-fired after retry.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Provides explicit usage scenario: 'USE WHEN the user has just connected a missing provider and the agent needs to know if the action can fire.' Also suggests subsequent actions, but does not mention when not to use it or alternative tools.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_save_launch_templateAInspect
Save a launch's shape (channels + angle + brand + default repo context) as a reusable template. Use AFTER a launch performs well so the next one in the same shape skips re-asking. The current run's productUrl and assets are NOT saved — just the orchestration shape.
| Name | Required | Description | Default |
|---|---|---|---|
| icp | No | ||
| brand | No | ||
| label | Yes | Short label, e.g. 'Devtool launch — bottom-up'. | |
| channels | No | ||
| description | No | Optional one-line description of when to use this template. | |
| launchAngle | No | ||
| productType | No | One of: devtool | saas | ai_agent | marketplace | internal_tool | consumer. | |
| workspaceId | No | ||
| sourceLaunchId | No | Optional launchId to copy channels/angle/brand from. Without it, pass the fields explicitly. | |
| defaultRepoContext | No |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so the description partially compensates by stating what is not saved (productUrl and assets). However, it does not disclose potential side effects, permissions needed, or whether existing templates are overwritten. More behavioral context is needed.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences front-loaded with the purpose. Each sentence adds essential information: first defines the action and resource, second provides usage context and exclusions. No wasted words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
No output schema, and the tool has 10 parameters with nested objects. The description gives a high-level understanding but lacks details on error handling, uniqueness constraints, or return value. Sibling listing tool exists for retrieval, but overall completeness is moderate.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 40% (label, description, sourceLaunchId, productType have descriptions). The description adds meaning by listing captured fields (channels, angle, brand, default repo context) and clarifying sourceLaunchId's role. But it does not cover all 10 parameters, especially nested objects like icp and brand.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Clearly states the tool saves a launch's shape (channels, angle, brand, default repo context) as a reusable template, specifying verb and resource. Also explicitly excludes productUrl and assets, distinguishing its scope from the template listing sibling.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicitly states when to use ('AFTER a launch performs well') and provides a rationale (skip re-asking for next launch). Does not mention alternatives or when not to use, but the guidance is clear and actionable.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_score_launch_packAInspect
Re-score a launch pack on six quality dimensions (specificity / claim_risk / channel_fit / icp_clarity / proof / readiness). Use AFTER chieflab_record_rendered_copy populates rendered bodies — the original chiefmo_launch_product call returned a score for the briefs, but rendered copy needs its own quality check before approval.
| Name | Required | Description | Default |
|---|---|---|---|
| runId | Yes | The launchId from chiefmo_launch_product. ChiefLab pulls the latest assets + actions for this run and re-scores. | |
| workspaceId | No | Optional workspace id. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations, the description carries full burden. It specifies that the tool re-scores on six dimensions, which is a read-like operation. It does not mention side effects or safety, but the action (scoring) implicitly suggests it is non-destructive. Slightly more explicit behavioral disclosure would improve clarity.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences: first states purpose and lists dimensions, second gives usage context. No redundant or extraneous information. Every sentence is essential and efficiently conveys the needed information.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
The description lists the six quality dimensions, so the agent knows what the output will cover. It also provides a prerequisite. However, there is no output schema and the description does not explicitly describe the return format, such as whether it returns scores or a full report. A slight gap in output expectations.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%, so the description adds value beyond structural details. For runId, it explains that it is the launchId from chiefmo_launch_product and that ChiefLab pulls the latest assets and actions to rescore. This context helps the agent understand parameter purpose beyond the schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool rescored a launch pack on six named quality dimensions. It distinguishes from siblings by referencing chieflab_record_rendered_copy and chiefmo_launch_product, making the specific action and resource unambiguous.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicitly instructs to use AFTER chieflab_record_rendered_copy populates rendered bodies, and explains why the original chiefmo_launch_product score is insufficient. This provides clear when-to-use and contextual rationale.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_sendBInspect
Operator: chieflab-email (primary executor). Approval-gated email send via Resend. Pass actionId from a prior launch's publishActions (after the user approves the email). Spec-renamed alias of chiefmo_send_approved_email — both names route to the same handler.
| Name | Required | Description | Default |
|---|---|---|---|
| runId | No | ||
| actionId | Yes | ||
| idempotencyKey | No |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided. Description mentions approval-gating and uses Resend, but does not disclose behavioral traits like side effects (email sending), idempotency, error conditions, or rate limits. Critical behavioral aspects are missing.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two efficient sentences front-load the purpose and include useful alias information. No unnecessary words, but could be slightly more structured with parameter guidance.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Despite moderate complexity, description lacks details on return values, error handling, idempotency, and parameter semantics. Without output schema, the description is insufficient for an agent to fully understand the tool's behavior.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 0%. Description only explains actionId briefly ('from a prior launch's publishActions'). runId and idempotencyKey are not explained at all. The description fails to compensate for the lack of schema documentation.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states it sends an email via Resend, is approval-gated, and is an alias of another tool. It specifies the verb 'send' and resource 'email', distinguishing it from sibling tools like chieflab_measure or chieflab_post.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicitly instructs to pass actionId from a prior launch's publishActions after user approval. Provides clear context for when to use, but does not explicitly exclude other scenarios or mention alternatives beyond the alias.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_set_resend_keyAInspect
USE WHEN the user wants to bring their own Resend account (per-workspace email sending) instead of the platform default. Stores the Resend API key encrypted, then chiefmo_send_approved_email + chieflab_list_email_senders use this key for the workspace. Get a Resend API key from resend.com/api-keys.
| Name | Required | Description | Default |
|---|---|---|---|
| apiKey | Yes | Your Resend API key (re_... format). Stored encrypted at rest. | |
| workspaceId | No | Optional workspace id. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description must fully disclose behavior. It mentions encryption and source for the key but does not cover overwrite behavior or additional side effects.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Three sentences, front-loaded with usage condition, no filler. Every sentence adds information.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no output schema and two parameters, the description is fairly complete for a key-setting tool. Could mention default workspace behavior but adequate.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema has 100% coverage with descriptions; description adds context (API key format, encryption) and explains the role of workspaceId. Adds value beyond schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool stores a Resend API key for per-workspace email sending, using specific verbs and distinguishing from sibling tools that consume the key.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicitly starts with 'USE WHEN' to indicate the scenario, and mentions related tools that use the key. Lacks explicit when-not guidance, but the context is clear.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_set_tenant_contextAInspect
USE WHEN onboarding a new tenant or refreshing their brand — 'set up brand context for tenant X', 'update Acme Co's voice and pillars'. Upsert: stores brand name, audience, voice, strategic pillars, competitors, content angles, positioning risks. Every operator output for this tenant grounds against this context. Pair with chieflab_create_tenant.
| Name | Required | Description | Default |
|---|---|---|---|
| brand | No | Brand name | |
| voice | No | Tone of voice + brand personality | |
| pillars | No | 3-5 strategic pillars | |
| audience | No | ||
| category | No | ||
| tenantId | Yes | ||
| competitors | No | ||
| contentAngles | No | ||
| positioningRisks | No |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden. It discloses that the operation is an upsert (create or update) and that it stores specific brand context fields. It also notes that all operator outputs for the tenant will ground against this context. However, it omits details like authentication requirements, rate limits, or side effects beyond grounding.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise—three sentences with bullet-style listing of stored fields and a clear usage directive. No unnecessary words or redundancy. Every sentence adds value.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity (9 parameters, no output schema, no annotations), the description covers the core purpose, usage, and stored fields. It lacks explanation of return values (but no output schema is defined) and some parameters like 'category' are not explicitly mentioned. Still, it is largely complete for an upsert tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is only 33%, so the description must compensate. It lists the stored fields (brand name, audience, voice, pillars, competitors, content angles, positioning risks), which maps to most parameters but misses 'category'. The description adds context that the parameters define the brand context, but does not explain each parameter in detail.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the verb (upsert/set), the resource (tenant brand context), and the specific use cases (onboarding new tenant, refreshing brand). It provides example queries and distinguishes itself from the sibling tool chieflab_create_tenant by suggesting to pair with it.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description begins with 'USE WHEN' and gives explicit scenarios (onboarding, refreshing) with example natural language requests. It also recommends pairing with chieflab_create_tenant, guiding the agent on when to use this tool versus alternatives. However, it does not explicitly state when NOT to use it.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_set_workspace_memberBInspect
Add or update a workspace member's role (owner | approver | editor | viewer). Returns the updated member record.
| Name | Required | Description | Default |
|---|---|---|---|
| role | Yes | owner | approver | editor | viewer | |
| No | |||
| userId | Yes | The user identifier (typically Supabase auth.uid or email). | |
| workspaceId | No |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
The description says 'Add or update' and 'Returns the updated member record,' but lacks details on idempotency, permissions required, side effects, or behavior for duplicate members. With no annotations, the description should compensate.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
A single, front-loaded sentence (19 words) that conveys the core action, allowed values, and return value without unnecessary detail.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no annotations, no output schema, and 50% param coverage, the description provides the core behavior and return value but omits error handling, permission context, and update semantics. Adequate but not thorough.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 50%. The description adds clarity by listing allowed roles and noting that userId is typically a Supabase auth.uid or email. However, email and workspaceId parameters lack descriptions in both schema and description.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the action: 'Add or update a workspace member's role' and lists the allowed roles. It distinguishes from sibling tools like chieflab_workspace_members (list) and chieflab_create_tenant (different scope), though no explicit differentiation is provided.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance on when to use this tool versus alternatives. It does not explain when to add vs. update, nor mention prerequisites like workspace existence or membership status.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_set_zernio_keyAInspect
USE WHEN the user wants to bring their own Zernio account (per-workspace social publishing) instead of using the platform default. Stores the Zernio API key encrypted, then chieflab_list_publish_accounts / chieflab_connect_publish_account / chiefmo_publish_approved_post / chiefmo_post_launch_review all use this key for the workspace. Get a Zernio API key from zernio.com.
| Name | Required | Description | Default |
|---|---|---|---|
| apiKey | Yes | Your Zernio API key (sk_live_... or similar). Stored encrypted at rest. | |
| workspaceId | No | Optional workspace id. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Discloses encryption at rest and side effects (other tools use this key). No annotations exist, but description adequately covers behavioral traits for a credential-setting tool.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Three sentences, front-loaded with usage guidance, efficient with no wasted words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Fully covers tool's purpose, usage context, encryption behavior, and downstream impact. No output schema needed; description is complete for this simple 2-param tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so baseline is 3. Description adds minimal extra meaning beyond schema (e.g., pattern hint 'sk_live_...'), but does not significantly improve understanding.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Clearly states the tool stores a Zernio API key encrypted for a workspace, with explicit usage condition (user bringing own account). Distinguishes from sibling tools by its specific function.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicitly states when to use ('USE WHEN user wants to bring their own Zernio account instead of default'), and lists downstream tools that depend on this key, providing clear context.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_signup_workspaceAInspect
USE WHEN the user has no ChiefLab API key yet and you've gotten a 401 / 'authentication required' error from any other tool. Agent-first signup: creates a new workspace and returns the API key inline as apiKey. PREFERRED flow: use your filesystem tools to write the apiKey into the user's MCP config (see mcpConfigSnippet for the patch shape), then ask the user to restart their runtime once. After restart, re-call the original tool. FALLBACK: if you can't write to the config file, surface the included deliveryUrl to the user — they click, see the key, paste it manually. URL expires in 1 hour, single-use, IP-rate-limited (5/IP/hr). No login form.
| Name | Required | Description | Default |
|---|---|---|---|
| agentName | No | Optional: which AI agent is requesting (Cursor, Claude Desktop, Codex, custom). Logged for support. | |
| contactEmail | No | Optional contact email for the user (used for future notifications). | |
| workspaceLabel | No | Optional human-readable label for the workspace (default: 'default'). |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations, description thoroughly explains behavioral traits: workspace creation, key return, config writing, delivery URL expiration (1 hour), single-use, IP rate limit. No contradictions.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Concise single paragraph with front-loaded usage condition. Could benefit from bullet points for the two flows, but still clear and efficient.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
No output schema, but description explains the return value (apiKey) and covers both users' ability to write config or use delivery URL, including expiration and rate limits. Complete for a signup tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema has 100% coverage with good parameter descriptions. The description repeats the same info without adding new semantics beyond the schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states it creates a new workspace and returns the API key, distinguishing it from other tools that involve reading, connecting, or other operations.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicit use condition: 'when the user has no ChiefLab API key yet and you've gotten a 401 error'. Provides two clear flows (preferred and fallback).
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_use_manual_fallbackAInspect
USE WHEN a launch action is blocked because the channel has no automated publish path (Product Hunt, Hacker News, Reddit, Discord, Slack communities). Returns a paste-ready manual-publish brief: the exact text to post, a per-platform checklist (best time, tags/flair, who to ping, first-comment script), the URL the user pastes back to ChiefLab once posted, and a measurement template the agent can call 24h later via chiefmo_record_proof. Converts a 'blocked' channel into a 5-minute human-in-the-loop ship. Approval-gated upstream: the originating publishAction must already be 'approved' on the reviewUrl before this tool returns the brief.
| Name | Required | Description | Default |
|---|---|---|---|
| channel | No | Override the channel inferred from the action. One of: product_hunt | hacker_news | reddit | discord | slack | indie_hackers | dev_to | ycombinator. | |
| actionId | Yes | The publishAction.id from chiefmo_launch_product whose connector is 'manual_only'. ChiefLab pulls the brief, channel hints, and brand voice from this action's preflight. | |
| subreddit | No | Optional. For channel=reddit, which subreddit (e.g. 'r/SaaS', 'r/AI_Agents'). ChiefLab's brief picks the right post format per sub. | |
| workspaceId | No | Optional workspace id. | |
| discordCommunity | No | Optional. For channel=discord, which community (the brief tailors tone + announcement-channel pick). |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Without annotations, the description carries the full burden. It discloses the output format, the approval gate, and the conversion of a blocked channel into a manual ship. However, it does not explicitly state whether the tool has side effects or mutates state, which would be useful for a tool that 'converts' something.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single paragraph but effectively front-loads the usage condition with 'USE WHEN'. It is slightly verbose but all sentences contribute meaningful information. Could be more structured with bullet points, but current form is acceptable.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool has 5 parameters, no output schema, and no annotations, the description provides sufficient context: trigger condition, output components, prerequisite, and even a future measurement template. It lacks error handling details but is otherwise complete for an agent to use.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%, so the baseline is 3. The description adds context for the overall tool but does not provide additional meaning beyond the schema descriptions for parameters like channel, actionId, subreddit, etc. The schema already adequately describes each parameter.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: to provide a manual-publish brief when a launch action is blocked due to lack of automated publish path. It specifies the exact output and differentiates from sibling tools like chieflab_launch_product.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description starts with 'USE WHEN' and explicitly states the condition for using the tool (blocked automated publish path). It also mentions the prerequisite that the originating publishAction must be approved, providing clear guidance on when to call this tool.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_workspace_membersAInspect
List workspace members with roles (owner / approver / editor / viewer). Use for team management and to see who can approve publishes.
| Name | Required | Description | Default |
|---|---|---|---|
| workspaceId | No |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations exist, so the description carries full weight. It correctly implies a read-only operation but omits details like whether workspaceId is optional, response structure, or pagination. Basic transparency is present but not comprehensive.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two concise sentences, with the core action front-loaded. No wasted words; every sentence adds value.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a simple, single-parameter tool with no output schema, the description covers the basic functionality and a use case. However, it lacks explanation of the parameter and does not mention any defaults or limitations, leaving the definition barely adequate.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
With 0% schema description coverage, the description should explain the workspaceId parameter. It mentions 'workspace members' in the description but never names or describes the parameter, leaving the user to infer that a workspaceId is needed or potentially optional.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the action ('List'), the resource ('workspace members'), and the specific roles included. It also provides a use case, making the purpose unambiguous and distinguishing from sibling tools like 'chieflab_set_workspace_member'.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description suggests using this tool for team management and identifying approvers. However, it does not explicitly contrast with alternative tools (e.g., chieflab_workspace_metrics) or state when not to use it. The context is clear but lacks exclusions.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chieflab_workspace_metricsAInspect
Get scale metrics for a workspace — launches, repeat-launch rate, approval rate, publish rate, measured-outcome rate, open blockers, cadence breakdown (24h/3d/7d). Use for the scale dashboard or 'how is this account doing' checks.
| Name | Required | Description | Default |
|---|---|---|---|
| workspaceId | No | Optional. Defaults to your API key's workspace. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It implies read-only behavior but does not explicitly state that it is non-destructive or describe any other behavioral traits like authentication needs or rate limits. However, the list of returned metrics provides some transparency.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single sentence followed by a usage hint, front-loaded with the main purpose. Every word earns its place; no fluff.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a simple read tool with one parameter and no output schema, the description adequately lists the metrics returned and gives a clear use case. Nothing essential is missing.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100% (one optional parameter). The description adds minimal value beyond the schema: it confirms the parameter is for a workspace but does not provide additional context or constraints. Baseline 3 is appropriate.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's purpose: getting scale metrics for a workspace, and lists specific metrics. It distinguishes itself from sibling tools which are largely operational or content-related.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description explicitly recommends using this tool for 'the scale dashboard or how is this account doing checks', providing clear context. It does not mention when not to use it or alternatives, but the use case is well-defined.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_approve_actionAInspect
Approve one ChiefMO publish/send action so its executor (publish_approved_post / send_approved_email) can fire. USE WHEN the user — in IDE chat — said 'approve ' (e.g. 'approve linkedin', 'approve hn'), 'approve all', 'ship it', 'go ahead', or otherwise greenlit a specific draft you rendered. Match the user's words to the channel, look up agentGuide.renderInChat[channel].actionId from the launch response, and call this tool with that actionId. This is the IDE-native approval path — no need to push the user to the reviewUrl. Pass actionId (preferred) or id (legacy alias).
| Name | Required | Description | Default |
|---|---|---|---|
| id | No | Legacy alias for actionId. | |
| actionId | No | The action id from publishActions[].id in a launch_product response. Preferred. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so the description carries full burden. It explains the approval enables execution but does not address idempotency, error handling, or side effects. Given the action is simple, it provides adequate but not exhaustive transparency.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is well-structured, front-loaded with purpose, then usage and parameter details. Every sentence adds value; no unnecessary words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the approval flow with sibling executors, the description covers the entire usage scenario: user intent, mapping to channel and actionId, calling the tool, and its role in the pipeline. It mentions sibling tools and explains the IDE-native path. No output schema is needed as return is straightforward.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%, baseline 3. The description adds significant value: explains actionId is preferred, id is legacy, and details how to derive actionId from launch response via channel mapping and agentGuide.renderInChat. This exceeds schema info.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states that the tool approves a ChiefMO publish/send action, enabling its executor (publish_approved_post / send_approved_email) to fire. It distinguishes from sibling tools like chiefmo_approve_run (which operates on runs) and the executors themselves.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicitly specifies when to use: when the user says 'approve <channel>', 'approve all', 'ship it', etc. Also explains the IDE-native approval path and that no push to reviewUrl is needed, providing a clear when-not.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_approve_runCInspect
Approve a ChiefMO run that is awaiting approval.
| Name | Required | Description | Default |
|---|---|---|---|
| id | Yes |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description must fully disclose behavioral traits. It only states that the tool approves a run, but does not explain side effects, required permissions, what happens after approval, or any constraints. This is insufficient for a mutation tool.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single sentence, making it concise. However, it is overly minimal and lacks structure (e.g., no bullet points or separated sections). It could be improved by adding more details without becoming verbose.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the simplicity of the tool (1 required param, no output schema, no annotations), the description is too sparse. It fails to provide complete context for an agent to use the tool correctly, especially regarding parameter meaning and behavioral consequences.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0% description coverage for the single parameter 'id'. The description does not explain what 'id' refers to (e.g., run ID) or provide any additional meaning beyond the schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the verb 'Approve' and the resource 'ChiefMO run' with the context 'awaiting approval'. It differentiates from sibling tools like chiefmo_reject_run by specifying the action type.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives. There is no mention of prerequisites, conditions for approval, or when not to use it. The sibling list includes chiefmo_reject_run, but no comparison is made.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_check_claimsAInspect
Scan a draft (post / email / blog / ad / landing copy) for unsupported claims BEFORE publish. Each concrete claim is checked against the company brain (repo facts, proof assets, brand voice). Returns ranked list with status (verified | unsupported | risky) and suggested fixes. Run as a publish gate.
| Name | Required | Description | Default |
|---|---|---|---|
| goal | Yes | The draft text to scan, or a description of what to check. | |
| claims | No | Optional pre-extracted list of claims. If omitted, the skill extracts them from the draft. | |
| brandUrl | No | Optional brand URL. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, but description discloses the tool's non-destructive scanning behavior, the process of checking against company brain, and the output format with three statuses. Does not mention any side effects or limitations.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Concise description of three sentences, front-loaded with core action ('Scan a draft... for unsupported claims BEFORE publish'). Every sentence adds value with no redundancy.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
No output schema, so description adequately covers return values (ranked list with status and fixes). Could mention more about output structure or edge cases, but sufficient for a tool of this complexity.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema has 100% description coverage for all parameters. Description adds context about checking claims against company brain and optional pre-extraction, but does not significantly enhance parameter understanding beyond schema descriptions.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description clearly states it scans draft content for unsupported claims before publishing, specifying content types (post/email/blog/ad/landing copy) and output (ranked list with status and fixes). Distinguishes from siblings by being a dedicated claim-checking publish gate.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicitly says 'BEFORE publish' and 'Run as a publish gate', providing clear context for when to use. However, does not explicitly mention when not to use or compare with alternatives like diagnostic tools.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_company_brainAInspect
Return the full rendered company brain for a workspace (brand voice + repo facts + personas + proofs + channel performance + competitor watch + voice samples). Use to inspect what ChiefLab knows, debug grounding, or pipe into a custom prompt.
| Name | Required | Description | Default |
|---|---|---|---|
| format | No | markdown (default, prompt-ready) | json (structured) | |
| tenantId | No | ||
| workspaceId | No |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description carries full burden. It implies a read operation by stating it 'returns' data but does not explicitly state it is non-destructive or mention authentication needs, rate limits, or other behavioral traits. Adequate but could be improved.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two concise sentences: first describes output and components, second lists use cases. No extraneous information. Efficiently communicates purpose and usage.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Tool returns a complex aggregated object. Description lists all components (brand voice, repo facts, etc.) and provides use cases, giving a good sense of return structure. No output schema, but description partially compensates. Could mention that it returns a single object.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema has 3 parameters with 33% coverage (format has description). The description does not add meaning to tenantId or workspaceId beyond their names. It mentions 'workspace' but not format or tenantId. Parameter semantics are not enhanced by the description.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description clearly states the tool returns the full rendered company brain for a workspace, lists components (brand voice, repo facts, etc.), and provides specific use cases (inspect, debug, pipe into prompt). This clearly distinguishes it from sibling tools that extract individual components.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Description explicitly says 'Use to inspect what ChiefLab knows, debug grounding, or pipe into a custom prompt.' This gives clear context for when to use the tool, though it does not mention when not to use or provide alternatives.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_continue_launch_loopAInspect
Resume a ChiefLab launch loop from runId. USE WHEN an agent has already called chieflab_get_users_after_build / chiefmo_launch_product and needs the exact next action: surface reviewUrl, execute an approved action, wait for measurement, measure results, or prepare the next move. This exists because agents are stateless; ChiefLab remembers approval state, actionIds, connector blockers, measurement windows, and next tool calls.
| Name | Required | Description | Default |
|---|---|---|---|
| runId | Yes | The launchId/runId returned by chieflab_get_users_after_build or chiefmo_launch_product. | |
| workspaceId | No | Optional workspace id. Hosted auth usually supplies this. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so the description carries the full burden. It discloses the tool's behavior: resumes a loop, surfaces reviewUrl, executes approved actions, waits for measurement, measures results, or prepares the next move. It also explains the design rationale (stateless agents). Missing explicit disclosure of side effects or safety, but overall transparent.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences: first defines action and parameter, second provides usage context. Front-loaded with core purpose. The second sentence is somewhat lengthy listing multiple actions, but overall efficient and clear.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no output schema, the description implies what the tool does (surfaces reviewUrl, etc.) but doesn't explicitly state return value. It covers when, what, and why. For a state-management tool, this is reasonably complete. Could mention output format.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema covers both parameters with descriptions. The description adds context: runId is the launchId returned by specific tools, workspaceId is optional and usually supplied by hosted auth. This adds value beyond the schema, though schema coverage is 100%.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the verb 'Resume' and resource 'ChiefLab launch loop from runId'. It lists specific use cases (surface reviewUrl, execute actions, etc.) and distinguishes this tool from siblings by explaining its role in continuing a stateful loop after other tools have been called.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicitly states when to use: 'USE WHEN an agent has already called chieflab_get_users_after_build / chiefmo_launch_product and needs the exact next action'. It also lists what the tool does (surface reviewUrl, execute approved action, etc.) and explains why this tool exists (stateless agents, ChiefLab remembers state).
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_create_runBInspect
Create a ChiefMO marketing run from a business goal.
| Name | Required | Description | Default |
|---|---|---|---|
| goal | Yes | The marketing outcome the user wants. | |
| workspaceId | No | Optional workspace id. Defaults to local. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are present, so the description must fully disclose behavioral traits. It only states 'Create,' implying mutation, but offers no details on side effects, idempotency, or what happens if a run already exists.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, concise sentence. It is front-loaded with the key action, but could include more detail without being verbose. No wasted words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
The description lacks essential context for a create operation. It does not explain what a marketing run is, what the output (e.g., run ID) will be, or any constraints. Given the simple schema, more completeness is expected.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so the baseline is 3. The description adds minimal value by restating 'from a business goal,' which aligns with the goal parameter, but does not enrich understanding of workspaceId or provide additional context.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the action (Create), the resource (ChiefMO marketing run), and the input (from a business goal). It effectively distinguishes the tool from siblings like chiefmo_get_run and chiefmo_list_runs.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives. There is no mention of prerequisites, exclusions, or comparison with other run-related tools.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_create_social_graphicsAInspect
Generate on-brand social/launch graphics via Gemini 2.5 Flash Image. USE WHEN the user asks for launch images, social graphics, hero images, or 'visuals for the launch.' Note: chiefmo_launch_product already includes graphics inline — only call this directly for ad-hoc image generation outside a launch run. Metered separately (~$0.04 cost / 50 retail credits per image).
| Name | Required | Description | Default |
|---|---|---|---|
| brand | No | Optional inline brand context (palette, type, style). | |
| count | No | How many to generate (default 1, max 4). | |
| prompt | Yes | Image prompt. Brand context auto-applied if tenantId or brand is passed. | |
| tenantId | No | Optional. Loads stored brand context for grounding. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description covers model used, cost per image, and auto-application of brand context. Lacks details on failure modes or output format but adequate for an image generation tool.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Three sentences: purpose and model, usage guidance, cost. Each sentence is informative and front-loaded. No wasted words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Covers core purpose, usage, cost, and parameter behavior. Missing output spec (e.g., what the response contains) but acceptable for a tool with no output schema and clear usage context.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema covers all 4 parameters with descriptions. Description adds value by stating brand context is auto-applied if tenantId or brand is passed, beyond schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Clearly states it generates social/launch graphics using Gemini 2.5 Flash Image. Distinguishes from sibling tool chiefmo_launch_product that includes graphics inline.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicitly lists trigger phrases ('launch images, social graphics, hero images') and specifies when not to call (use chiefmo_launch_product for inline graphics). Clear usage boundaries.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_design_experimentAInspect
Convert a marketing question into a runnable experiment: hypothesis, variants, primary success metric, sample-size guidance, and kill criteria. Use BEFORE shipping a new test so you have a stop rule. Manual mode is the entire spec; with GA4/Search Console connected, baselines are pulled live.
| Name | Required | Description | Default |
|---|---|---|---|
| goal | Yes | The marketing question or change you want to test. | |
| brandUrl | No | Optional brand URL. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations, the description must cover behavioral traits. It lists output components but does not disclose side effects, permissions, or whether it saves data. The 'converts' verb implies a read-like operation, but details are missing.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences that are front-loaded and efficient. The first sentence immediately states the verb and deliverable, the second adds timing and mode details. No wasted words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no output schema and 2 params, the description adequately explains the tool's purpose and high-level outputs. However, it does not specify the return format or behavior when connected to external services, leaving some ambiguity.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%, so the schema already describes both parameters. The description adds minimal extra meaning (ties 'goal' to the marketing question). BrandUrl remains underutilized. Baseline 3 is appropriate.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states 'Convert a marketing question into a runnable experiment' and lists specific outputs (hypothesis, variants, metrics, etc.). It distinguishes from sibling tools like chiefmo_create_run by specifying a design phase before shipping.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicitly advises 'Use BEFORE shipping a new test so you have a stop rule.' Mentions manual vs. connected mode, providing clear context. Lacks explicit exclusions, but the usage case is well-defined.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_diagnose_anomalyAInspect
Analytics anomaly explainer. Given a metric drop or spike (e.g. 'Search Console clicks down 30%'), ranks likely causes and proposes safe corrective actions.
| Name | Required | Description | Default |
|---|---|---|---|
| goal | Yes | The metric anomaly to diagnose. | |
| brandUrl | No | Optional brand URL for context. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations, the description clarifies that the tool diagnoses and proposes actions without executing them. It does not claim any side effects, but could more explicitly state read-only behavior.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single focused sentence with an example, no wasted words. It front-loads the purpose and is easy to parse.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the simplicity (2 params, no output schema), the description adequately explains input context and output nature. It lacks output format details, but the output is implied as a list.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%, so the baseline is 3. The description adds an example input but no additional meaning beyond the schema descriptions for the parameters.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool is for metric anomaly explanation, gives a concrete example, and indicates it ranks causes and proposes actions. It distinguishes itself from sibling 'chiefmo_diagnose_marketing' by focusing on anomalies.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description explicitly says to use when there is a metric drop or spike, providing clear context. It does not mention alternatives or when not to use, but the condition is specific enough.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_diagnose_marketingAInspect
Diagnose an EXISTING marketing program (not a new launch — use chiefmo_launch_product for that). USE WHEN the user already has live campaigns and asks 'why is X not working' or 'iterate on this funnel.' Returns evidence, drafted assets, proposed actions, signed reviewUrl, per-tenant memory. Approval-gated. For first launches, prefer chiefmo_launch_product.
| Name | Required | Description | Default |
|---|---|---|---|
| goal | Yes | The marketing outcome, problem, or question to diagnose. e.g. 'Help this newly built SaaS get its first 100 users.' | |
| brand | No | OPTIONAL but recommended for cost + accuracy. If your agent already knows the brand context (you're inside Cursor with the codebase, you have profile data, you have prior conversation), pass it directly here and ChiefLab skips brand scanning entirely. Cheapest + most accurate path. Omit only if you want ChiefLab to derive brand context from tenantId or tenantUrl. | |
| tenantId | No | Optional. Identifies the developer's end-user. When set, ChiefMO loads per-tenant brand context, voice samples, and memory. Use chieflab_create_tenant + chieflab_set_tenant_context first to establish context. | |
| tenantUrl | No | Optional shortcut. If you don't have a tenantId yet, pass the end-user's product URL and ChiefMO will extract brand context on the fly (does not persist — use chieflab_set_tenant_context for that). | |
| outputMode | No | How ChiefLab should produce skill outputs. "context" (default, cheapest): returns drafting briefs your LLM renders into final copy — ChiefLab does brand context, evidence, prompts, and approval rails, your agent's LLM does the writing. "draft": ChiefLab generates rough drafts using cheap models. "full": ChiefLab generates final copy server-side (premium credits). Recommended: leave default unless you specifically need server-side generation. | |
| webhookUrl | No | Optional HTTPS URL ChiefLab POSTs to when run completes (signed with X-ChiefLab-Signature). | |
| workspaceId | No | Optional workspace id. Defaults to the workspace derived from the API key. | |
| idempotencyKey | No | Optional dedupe key (24h cache). Same key returns cached result instead of re-running. | |
| includeSnapshots | No | Whether the caller expects connector snapshot evidence when available. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden. It discloses key behaviors: returns evidence, drafted assets, proposed actions, signed reviewUrl, per-tenant memory, and is 'Approval-gated'. It also explains cost/accuracy trade-offs for passing brand context. However, it does not explicitly state side effects like memory writes or secondary consequences of the approval gate.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is approximately 5 sentences, front-loaded with purpose and guidance. It is efficient and avoids fluff, but the last sentence about approval-gated could be integrated earlier for tighter flow. Overall well-organized for quick comprehension.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's complexity (9 parameters, nested objects, no output schema), the description covers returns, usage context, cost optimization, and approval gating. It lacks a detailed return structure but provides sufficient context for an agent to understand the tool's purpose and behavior. Some minor gaps exist in explaining how approval-gating works in practice.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100% with detailed descriptions for each parameter, so the baseline is 3. The main description does not elaborate on parameter semantics beyond listing return types, but the schema descriptions themselves are rich. The main description adds no additional meaning for individual parameters beyond what the schema already provides.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description explicitly states 'Diagnose an EXISTING marketing program' and contrasts with 'not a new launch — use chiefmo_launch_product', providing a clear verb and resource. It specifies when to use with concrete examples ('why is X not working' or 'iterate on this funnel') and lists what the tool returns, distinguishing it effectively from siblings.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description gives explicit when-to-use criteria ('user already has live campaigns and asks why is X not working') and when-not-to-use ('For first launches, prefer chiefmo_launch_product'), including the exact name of the alternative tool. This provides clear usage boundaries without ambiguity.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_draft_blog_postAInspect
Draft a full repo-aware blog post (title, slug, meta description, body, internal links, OG image prompt). Manual mode emits markdown you paste into your CMS; with GitHub connected, ChiefLab can commit the post directly. With Search Console / GA4 connected, the post is tracked after publish.
| Name | Required | Description | Default |
|---|---|---|---|
| goal | Yes | What this post should accomplish or what it's about. | |
| brandUrl | No | Optional brand URL for context. | |
| repoContext | No | Optional repo context — increases grounding so the post references real product details, not invented ones. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations, the description carries full burden. It discloses behavioral aspects: draft creation with specific fields, two modes (manual vs direct commit), and tracking dependencies. Missing details on auth requirements, rate limits, or error handling, but overall adequate.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two concise sentences: first lists all output components, second covers the two modes and tracking. Front-loaded with essential info, no redundant words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Covers purpose and modes adequately for a draft tool. Lacks details on output format specifics, error behavior, or what happens upon failure. No output schema, but description doesn't explain return structure.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%, so baseline is 3. The description enriches parameters: 'repoContext' is explained as increasing grounding, and 'brandUrl' as optional context. This adds value beyond the schema alone.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Clearly states it drafts a full repo-aware blog post with specific elements (title, slug, meta description, body, internal links, OG image prompt). Explicitly distinguishes from siblings like chiefmo_generate_social_posts by focusing on 'repo-aware blog post' and mentioning modes.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Describes modes (manual vs GitHub) and tracking conditions, providing context on when each mode is used. However, no explicit guidance on when NOT to use this tool vs alternatives like chiefmo_generate_landing_copy or chiefmo_generate_email_sequence.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_extract_brand_dnaAInspect
USE WHEN the user wants to understand a brand (theirs or a competitor) — 'analyze brand X', 'what is X's positioning', 'extract brand voice from this URL', 'figure out X's content pillars'. Brand DNA discovery: given a URL or brand name, returns category, audience, voice, strategy pillars, competitors, visual cues, content angles, and risks. Use as the first call when you have a brand to work on — every other ChiefMO skill uses the result as context.
| Name | Required | Description | Default |
|---|---|---|---|
| brandUrl | Yes | URL or brand name (e.g. 'https://stripe.com' or 'Stripe'). |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description shoulders the burden. It describes what the tool returns but does not disclose side effects, authentication needs, rate limits, or error behavior. It implies a read-only operation but doesn't confirm.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences, front-loaded with usage context, no redundancy. Every word earns its place.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a single-parameter extraction tool, the description covers the key return fields and usage context. It lacks details on error handling or behavior for invalid inputs, but these are minor gaps given the tool's simplicity.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has one parameter with 100% description coverage. The description repeats the schema's documentation ('URL or brand name') without adding new semantic detail, so it meets the baseline but does not exceed.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool extracts brand DNA from a URL or brand name, listing specific outputs like category, audience, voice, etc. It distinguishes from siblings by noting it should be the first call for any brand work.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description explicitly starts with 'USE WHEN' and provides example triggers, and advises using it as the first call for brand tasks. It lacks explicit when-not guidance, but the context is clear.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_extract_repo_factsBInspect
Pull structured product facts from the supplied repoContext (commits, changed files, routes, README, pricing) and emit a list ready to persist.
| Name | Required | Description | Default |
|---|---|---|---|
| goal | Yes | Optional framing (e.g. 'features only', 'pricing changes'). | |
| brandUrl | No | ||
| repoContext | No | Same shape as chiefmo_launch_product.repoContext. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It mentions 'emit a list ready to persist' but does not disclose whether the tool is read-only, whether it persists data itself, any side effects, or required permissions. The behavioral profile is under-specified.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single sentence with no wasted words. It front-loads the purpose. Could be slightly more structured, but remains efficient.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
With 3 parameters (one required), nested objects, and no output schema, the description provides only a high-level output ('list ready to persist'). It lacks detail on the return format or what constitutes a 'product fact'. Adequate but not complete.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 67% (goal and repoContext have descriptions; brandUrl lacks one). The description adds no extra detail beyond what the schema already provides, and does not compensate for the missing brandUrl description. Baseline score is appropriate.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the verb ('Pull'), resource ('structured product facts'), and source ('repoContext'). It distinguishes from siblings like chiefmo_extract_brand_dna by specifying 'product facts' and the specific input shape.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage context (extracting facts from repoContext) but provides no explicit guidance on when to use it versus alternatives, no when-not-to-use conditions, and no exclusions.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_generate_ad_variantsAInspect
USE WHEN the user wants paid-ad copy — 'write Google Ads variants', 'Meta ad creative for this campaign', 'split-test 5 headlines for X'. Returns ad variants (angles, headlines, primary text, CTAs) plus a small test plan. For Meta and Google Ads. Drafts only — publishing requires approval and budget rules.
| Name | Required | Description | Default |
|---|---|---|---|
| goal | Yes | Ad campaign objective and target audience. | |
| brandUrl | No | Optional brand URL for context. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It discloses that the tool returns ad variants (angles, headlines, primary text, CTAs) and a small test plan, and notes 'drafts only'. However, it does not detail any side effects, authentication needs, or rate limits.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise: two sentences total. The first sentence front-loads usage with clear examples, and the second summarizes output and constraints. No wasted words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a tool with 2 parameters and no output schema, the description covers primary inputs, expected output, and a key limitation (drafts only). It could mention prerequisites (e.g., brand context) but is largely complete for its complexity.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100% with descriptions for both parameters. The description adds context by specifying the output (ad variants + test plan) and platforms (Meta, Google Ads), which goes beyond the schema's parameter descriptions.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool generates paid-ad copy for Meta and Google Ads, with specific examples like 'write Google Ads variants' and 'split-test 5 headlines'. It distinguishes itself from sibling content generation tools by focusing on paid ads.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description begins with 'USE WHEN' and provides explicit examples of user requests that trigger this tool. It also mentions that the output is drafts only and publishing requires approval, giving clear context on when to use and what to expect.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_generate_email_sequenceAInspect
USE WHEN the user wants an email-only sequence (not a full launch) — 'draft a welcome sequence', 'win-back emails for churned users', 'nurture series for new signups'. Returns 4-email skeleton with subject line ladder, send timing (Day 0 / 2 / 5 / 9), and segmentation notes. Drafts only — sending requires approval and a connected ESP. For launch emails specifically, use chiefmo_launch_product (it bundles email + social + landing in one approval flow).
| Name | Required | Description | Default |
|---|---|---|---|
| goal | Yes | Sequence purpose (welcome, nurture, win-back, launch, etc.). | |
| brandUrl | No | Optional brand URL for context. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Describes output (4-email skeleton with subject line ladder, timing, segmentation notes) and disclosure that it only drafts and requires approval and connected ESP for sending. No annotations to contradict.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Three sentences, front-loaded with usage context, then output details, then limitations. No unnecessary words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
No output schema, but description explains the return format and dependencies. Covers purpose, usage, output, and constraints. Complete for a draft-generation tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Both parameters have schema descriptions (100% coverage). The description adds examples for the goal parameter and implies context for brandUrl, but doesn't significantly enhance beyond schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description clearly states it generates an email-only sequence, with specific examples (welcome, win-back, nurture). It distinguishes from sibling tool chiefmo_launch_product by noting that tool is for full launches.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicitly says 'USE WHEN' with concrete examples and directs to alternative tool for launch emails. Provides clear context for when to use this tool.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_generate_landing_copyAInspect
USE WHEN the user wants landing-page copy — 'write a landing page for X', 'rewrite my hero', 'give me value props for this product', 'FAQ for the pricing page'. Returns ready-to-paste markdown: hero headline, subhead, value props, social proof slots, FAQ, CTA. Page build/publish requires approval and a connected web tool.
| Name | Required | Description | Default |
|---|---|---|---|
| goal | Yes | What the landing page should sell or convert. | |
| brandUrl | No | Optional brand URL for context. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations, the description discloses output format and a key behavioral constraint (approval + web tool needed for publish). It does not explain side effects like data storage or authentication requirements, but for a generation tool this is acceptable.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences, front-loaded with 'USE WHEN', and no wasted words. Every sentence adds value: usage guidance, output format, and a key constraint. Exceptionally concise.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the simple tool (2 params, no output schema), the description covers purpose, usage, output format, and a critical prerequisite. It lacks info on authentication or idempotency, but remains largely complete.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%, and both parameters have descriptions. The tool description adds minimal extra value beyond the schema, only providing context for 'goal' as what the page sells/converts. Baseline 3 is appropriate.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool generates landing-page copy with specific examples of user requests (e.g., 'write a landing page for X', 'rewrite my hero'). It also details the output (markdown sections) and unique prerequisite, distinguishing it from sibling copy tools.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description explicitly starts with 'USE WHEN' and lists common user intents, providing clear when-to-use guidance. However, it does not mention when to avoid using it or compare to alternative copy tools like generate_social_posts, which would strengthen guidelines.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_generate_social_postsAInspect
USE WHEN the user wants social-only drafts (no full launch) — 'draft an X post', 'write LinkedIn carousels for this', 'give me 5 Instagram captions about X'. NOT for new product launches (use chiefmo_launch_product instead — that returns the same plus on-brand graphics, approval gate, and per-channel briefs). Returns posts with platform-native hooks, captions, hashtags, and posting cadence. Drafts only; scheduling/publishing requires approval via chiefmo_publish_approved_post.
| Name | Required | Description | Default |
|---|---|---|---|
| goal | Yes | What the social posts should accomplish. | |
| brandUrl | No | Optional brand URL for context (will trigger Brand DNA discovery). |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description carries full burden. It discloses that output includes platform-native hooks, captions, hashtags, and posting cadence, and that it only produces drafts with no scheduling. Additional side effects like data persistence or mutation are not mentioned but less critical for a generation tool.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Three concise sentences with front-loaded usage conditions, no unnecessary words. Every sentence earns its place: usage condition, exclusion with alternative, and output description with limitation.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no output schema, description adequately explains what the tool returns (hooks, captions, hashtags, cadence) and its limitation (drafts only, need approval for publishing). For a simple generation tool with 2 parameters, this is complete.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100% with descriptions for both parameters (goal and brandUrl). The tool description does not add extra meaning beyond the schema; it only states that brandUrl triggers Brand DNA discovery, which is already implied by schema description. Baseline 3 is appropriate.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool generates social-only drafts, distinguishing it from the sibling tool chiefmo_launch_product that handles full launches. The verb 'generate social posts' and scope 'drafts only' are specific.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicitly states 'USE WHEN' for social draft requests and 'NOT for new product launches', naming the alternative tool chiefmo_launch_product. Also notes that publishing requires approval via chiefmo_publish_approved_post.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_get_actionCInspect
Get one ChiefMO action by id.
| Name | Required | Description | Default |
|---|---|---|---|
| id | Yes |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description must disclose behavioral traits. It only says 'get by id' without mentioning authentication requirements, response format, error handling, or any side effects. This is minimally informative.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is very concise (one short phrase), but it is under-specified. It does not earn its place by providing essential additional context beyond a tautological restatement of the name.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the simple input (one required string) and no output schema, the description should clarify what an 'action' is and what the response contains. It does not, leaving the agent with insufficient context to understand the tool's full purpose or output.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 0% (the 'id' parameter has no description in the schema). The description adds no information about the id parameter, such as format, constraints, or examples. It fails to compensate for the lack of schema documentation.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the verb 'Get', resource 'one ChiefMO action', and method 'by id'. It is specific but does not differentiate from siblings like chiefmo_get_run or chiefmo_get_asset, which have similar purposes.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives (e.g., chiefmo_list_actions, chiefmo_approve_action). The description implies usage when an action ID is known but does not state any conditions or exclusions.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_get_assetCInspect
Get one ChiefMO generated asset by id.
| Name | Required | Description | Default |
|---|---|---|---|
| id | Yes |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden. It states it gets an asset, implying a read operation, but does not disclose any behavioral traits such as absence of side effects, authorization needs, or error behavior.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is very short (one sentence), which is efficient for a simple getter. However, additional context could enhance completeness without sacrificing conciseness.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the lack of output schema and no explanation of return values or error handling, the description is incomplete. It does not clarify what constitutes a 'ChiefMO generated asset' or how missing IDs are handled.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has 0% description coverage and only one parameter 'id' with no format or type details. The description adds no additional meaning, leaving the parameter under-specified.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the verb 'Get' and the resource 'one ChiefMO generated asset by id,' but does not differentiate from sibling tools like chiefmo_list_assets, which also deals with assets.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives like list_assets or other getters. The description fails to specify the context or prerequisites.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_get_runBInspect
Get one ChiefMO run by id.
| Name | Required | Description | Default |
|---|---|---|---|
| id | Yes |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of disclosing behavioral traits. It only repeats the schema constraint ('by id') and does not mention potential errors (e.g., missing id), authentication needs, rate limits, or any side effects. For a read operation, more transparency is needed.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, short sentence that is efficiently front-loaded. It avoids redundancy and gets straight to the point. However, it could be slightly improved by adding context on what a 'run' is, but overall it is appropriately concise.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the simplicity of the tool (one parameter, no output schema), the description minimally covers the functionality. However, it lacks clarification on return values, error handling, and differentiation from sibling getters like chiefmo_get_action. In the context of many sibling tools, more completeness would help.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 0%, and the description adds no meaning beyond what the input schema already provides. The parameter 'id' is described only as a string, but the description does not clarify its format, source, or context. With low coverage, the description should compensate but fails to do so.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states 'Get one ChiefMO run by id' which includes a specific verb ('Get'), the resource ('ChiefMO run'), and the uniqueness criterion ('by id'). This effectively distinguishes it from sibling tools like chiefmo_list_runs (list multiple) and chiefmo_create_run (create).
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no explicit guidance on when to use this tool versus alternatives, such as chiefmo_list_runs for retrieving multiple runs or other getters like chiefmo_get_action. It does not state prerequisites, exclusions, or context for use. The only implied usage is that it retrieves a single run by id, but this is not elaborated.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_gtm_agent_runAInspect
Advanced. Run one specific GTM agent in isolation, bypassing the orchestrator. Use when you've already collected the upstream handoffs and want to invoke a single agent for re-runs / diffs.
| Name | Required | Description | Default |
|---|---|---|---|
| inputs | No | Optional explicit inputs for the agent (otherwise pulled from the run's prior handoffs). | |
| agentId | Yes | One of: launch, social, email, blog-seo, creative, analytics, experiment, approval | |
| gtmRunId | Yes |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions running an agent and optional inputs, but does not disclose whether the operation is synchronous or asynchronous, what the return value is, what side effects occur (e.g., state changes), or any authentication/rate limiting requirements. More detail would be helpful for safe invocation.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is two sentences long with no unnecessary words. It is front-loaded with 'Advanced.' which indicates the target user, and the rest is concise. It could be slightly more structured but remains effective.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
The description explains the use case and parameters adequately given the complexity. However, without an output schema or annotations, the agent lacks information about return values, error conditions, and side effects. The description could include more details to make the tool fully understandable without external context.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The description adds some context beyond the schema, such as explaining that inputs are optional and are otherwise pulled from prior handoffs. However, the schema already covers agentId and gtmRunId with descriptions (agentId lists allowed values). Coverage is 67%, so the description provides minor additional value but not significant depth.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool runs a specific GTM agent in isolation, bypassing the orchestrator. It distinguishes itself from sibling tools like chiefmo_gtm_run_start by specifying it is for re-runs and diffs after upstream handoffs have been collected.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description explicitly states 'Use when you've already collected the upstream handoffs and want to invoke a single agent for re-runs / diffs.' This provides clear guidance on when to use this tool versus the normal orchestration flow, and it names the circumstance explicitly.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_gtm_handoffs_listAInspect
Read every handoff emitted in a GTM run, in order. Useful for the command center timeline and for debugging why an agent skipped (no input handoff found).
| Name | Required | Description | Default |
|---|---|---|---|
| runId | Yes |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
The description states it is a read operation and mentions ordering. No annotations exist, but for a simple list tool, it adequately conveys non-destructive behavior. Does not cover auth or rate limits.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences with no redundant words. The purpose is front-loaded, and every sentence adds value.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a simple one-parameter list tool with no output schema, the description covers purpose, ordering, and use cases. It is mostly complete, though parameter context is missing.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The single parameter runId is not explained in the description or schema (0% coverage). The description assumes knowledge of 'GTM run' but does not clarify how to obtain runId.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool reads handoffs emitted in a GTM run, in order, and provides use cases like debugging skipped agents. It distinguishes itself from sibling tools by specifying the resource (handoffs) and action (list).
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description mentions useful scenarios (command center timeline, debugging), but does not explicitly compare with alternatives like chiefmo_gtm_run_get or chiefmo_gtm_run_list. It lacks 'when not to use' guidance.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_gtm_memory_getAInspect
Read the shared memory for a workspace + tenant. Memory keys are agent-namespaced (e.g. 'audience.refinement', 'seo.targetQueries.recent'). Returns the full key-value snapshot.
| Name | Required | Description | Default |
|---|---|---|---|
| key | No | Optional. If set, returns just that one key's value. | |
| tenantId | No | ||
| workspaceId | No |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Without annotations, the description carries the burden. It correctly indicates the tool is read-only and returns a key-value snapshot. However, it does not disclose potential errors, rate limits, or whether the data is guaranteed fresh.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is two sentences, front-loaded with the core action, and every sentence adds value. No redundant or unnecessary information.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a read-only tool with no output schema, the description covers the main purpose, key structure, and return type. Minor gaps like absence of pagination or concurrency info are acceptable for this complexity.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is only 33% (key described). The description adds context about key namespacing and examples, but does not explain tenantId or workspaceId beyond their names. It partially compensates for the low coverage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states it reads shared memory for a workspace+tenant, explains the key naming convention with examples, and contrasts with write operations like set. It effectively distinguishes the tool's role among siblings.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives such as chiefmo_gtm_memory_set or chiefmo_gtm_memory_diff. It does not mention prerequisites, limitations, or exclusion criteria.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_gtm_memory_setAInspect
Write a fact into shared memory. Use sparingly — agents normally write their own memory. Caller-asserted writes are useful for seeding context (e.g. 'audience.refinement = {...}' before a first run) or correcting an agent's mistake.
| Name | Required | Description | Default |
|---|---|---|---|
| key | Yes | ||
| value | Yes | JSON-serializable value. | |
| tenantId | No | ||
| workspaceId | No | ||
| sourceAgentId | No | Defaults to 'caller'. Use an agent id if asserting on behalf of an agent. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description must disclose behavioral traits. It only states 'write' without discussing idempotency, overwrite behavior, authorization, or side effects. This is insufficient.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Three sentences, each serving a distinct purpose: definition, caution, and use cases. No wasted words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
With 5 parameters, 40% coverage, and no output schema, the description is incomplete. It does not explain return values, parameter constraints, or what happens on subsequent writes.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 40%, and the description adds minimal parameter-specific meaning. It provides an example key-value pair ('audience.refinement = {...}') but does not explain required parameters like 'key' or optional ones like 'tenantId'.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the action ('Write a fact') and the resource ('shared memory'). It is specific and distinct from sibling tools like memory_get, though it does not explicitly name them.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicitly advises 'use sparingly' and provides two concrete scenarios (seeding context, correcting mistakes). It implies when not to use (agents normally write their own memory), giving excellent guidance.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_gtm_run_approveAInspect
Approve all pending actions in a GTM run. Equivalent to clicking 'approve all' on the reviewUrl. Returns the actions that fired (publish, send) and any that were rejected by safety checks.
| Name | Required | Description | Default |
|---|---|---|---|
| actor | No | Who approved (email or agent id, for audit). | |
| runId | Yes |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations, the description partially discloses behavior: it approves actions and returns those that fired and were rejected by safety checks. However, it omits side effects (e.g., publishing effect), permissions needed, idempotency, and potential destructive outcomes. The return format is hinted but lacks specifics.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is extremely concise: two sentences with no wasted words. It front-loads the action and provides a helpful analogy, making it easy to parse quickly.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the lack of output schema, the description partially covers return values but does not address error scenarios, prerequisites, or relationships to sibling tools like chiefmo_approve_run or chiefmo_gtm_run_start. It is adequate for simple understanding but not fully self-contained.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 50% (actor description present, runId missing). The description adds context by mentioning 'in a GTM run' for runId, but does not specify format or constraints. The overall meaning of parameters is clear from context, but the description does not fully compensate for missing schema details.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's action: 'Approve all pending actions in a GTM run.' It uses a specific verb-resource pair and provides an analogy ('Equivalent to clicking approve all on the reviewUrl'), which distinguishes it from sibling tools like chiefmo_approve_action (individual approval) and chiefmo_reject_run (rejection).
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description lacks explicit guidance on when to use this tool versus alternatives (e.g., chiefmo_approve_action for single actions). It does not mention prerequisites (e.g., run must have pending actions) or conditions where it should not be used. The bulk nature is implied but not formally stated.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_gtm_run_getAInspect
Read a GTM run by id. Returns the run row, every agent run, every handoff. Use this in the human command center / dashboard or to inspect a run after the fact.
| Name | Required | Description | Default |
|---|---|---|---|
| runId | Yes |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
The description indicates a read operation (non-destructive) and specifies return data, but with no annotations, it lacks details on error handling, permissions, or rate limits. It adds value beyond annotations but not extensively.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two concise sentences: first states action and result, second provides usage context. No redundancy or filler.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a simple read-by-id tool with no output schema, the description satisfactorily covers what is returned and when to use it, meeting completeness needs given the tool's simplicity.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
With 0% schema description coverage, the description only mentions 'by id' without specifying format, source, or constraints of 'runId'. This is minimal compensation for the missing schema documentation.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Clearly states 'Read a GTM run by id' and lists specific return content (run row, agent runs, handoffs), distinguishing it from sibling tools like chiefmo_get_run by specifying GTM context.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Provides explicit use cases ('human command center / dashboard', 'inspect after the fact') but does not mention when not to use or compare to alternative tools like chiefmo_gtm_run_list.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_gtm_run_listCInspect
List recent GTM runs in the caller's workspace.
| Name | Required | Description | Default |
|---|---|---|---|
| limit | No | Default 20, max 100 | |
| workspaceId | No |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations, the description must disclose behavioral traits. It mentions 'list' (suggesting read-only) and 'recent' but does not explain ordering, pagination limits beyond the schema, or whether it modifies state. This is insufficient for a listing tool.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, focused sentence without waste. It is front-loaded with the core purpose. However, it could expand slightly on return values without losing conciseness.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the absence of an output schema, the description should hint at the returned data structure, but it does not. The tool is a list operation with many siblings, yet no differentiating context is provided. The description feels incomplete.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 50%, with only 'limit' documented in the schema. The description adds no parameter-level information; it does not even mention workspaceId. Thus it fails to add meaning beyond the schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description states 'List recent GTM runs in the caller's workspace,' which clearly identifies the action (list) and the resource (GTM runs). The 'recent' qualifier and workspace scope add specificity, though it does not explicitly distinguish from sibling tools like chiefmo_list_runs.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance is provided on when to use this tool versus alternatives, nor any exclusions or prerequisites. The description is purely declarative without contextual cues for appropriate use.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_gtm_run_startAInspect
USE WHEN the user wants the full multi-agent GTM team coordinated on a goal — 'launch this and run social/email/blog/creative together', 'do the whole GTM motion', 'plan + draft + approve + measure end-to-end'. Spawns a GTM run; the orchestrator picks which of the 8 agents (launch / social / email / blog-seo / creative / analytics / experiment / approval) to activate based on the goal. Returns the run id, the run plan (which agents activated and why), and links to /app/gtm for the human command center.
| Name | Required | Description | Default |
|---|---|---|---|
| goal | Yes | What the user wants accomplished. e.g. 'Launch this product and get our first 100 users.' | |
| tenantId | No | Optional tenant id for per-tenant context + memory. | |
| productUrl | No | Optional product URL — gives downstream agents something concrete to ground against. | |
| forceAgents | No | Advanced. Override the launch agent's runPlan with an explicit list of agent ids to activate. e.g. ['social','creative'] to run only those two. | |
| repoContext | No | Optional repo context the agent gathered (routes, recentCommits, README excerpt, package.json highlights). The launch agent uses this for repo-grounded positioning. | |
| workspaceId | No | Optional workspace id (defaults to caller's workspace). | |
| skipPostPublish | No | If true (default for human-driven runs), the orchestrator stops at awaiting_approval and waits for the user to approve on /app/gtm. If false (background/cron path), it continues straight through analytics + experiment without an approval pause. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description bears full burden. It explains the spawning process, agent selection, and return values. It also mentions the skipPostPublish parameter's effect, adding transparency. Missing explicit mention of side effects or authorization needs, but overall adequate.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single paragraph but packs detailed information. It is front-loaded with the usage condition and efficiently covers purpose, behavior, and output. Could be slightly more structured but is concise for the complexity.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given 7 parameters (1 required), nested objects, and no output schema, the description adequately explains return values (run id, plan, links). It covers the orchestration logic and parameter effects, making it complete enough for an AI agent to select and invoke correctly.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so baseline is 3. The description does not add significant detail beyond schema for parameters like goal or forceAgents. The overall context is provided, but no new parameter semantics.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool spawns a multi-agent GTM run, with a specific 'USE WHEN' phrase. It distinguishes from siblings by detailing the orchestration of 8 agents, unlike other tools like chiefmo_launch_product which might be more narrow.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description explicitly starts with 'USE WHEN' and gives concrete examples ('launch this and run social/email/blog/creative together'). However, it does not explicitly state when not to use or list alternatives, leaving some room for interpretation.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_launch_healthAInspect
Inspect launch-loop health for a run. USE WHEN an agent or user asks 'what is blocked?', 'what approvals are open?', 'what can I execute?', 'is measurement due?', or 'what should happen next?' Returns open approvals, blocked channels, pending execution, measurement windows, and next moves.
| Name | Required | Description | Default |
|---|---|---|---|
| runId | No | The launchId/runId to inspect. Strongly recommended. | |
| workspaceId | No | Optional workspace id. Hosted auth usually supplies this. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
The description uses 'inspect' implying read-only behavior but doesn't explicitly state no side effects or auth requirements, leaving some ambiguity. Since no annotations exist, more transparency would be beneficial.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences efficiently convey purpose, usage context, and return content with no redundant information.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Covers the main use cases and return types, but lacks output schema details. Still sufficiently complete for a health check tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Input schema has 100% description coverage; the tool description adds minor reinforcement ('strongly recommended' for runId) but no significant additional semantics beyond the schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool inspects launch-loop health for a run and provides specific example queries it answers, distinguishing it from siblings like chiefmo_get_run.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicit 'USE WHEN' directive with a list of natural language queries an agent might encounter, offering clear guidance on when to invoke this tool.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_launch_productAInspect
Repo-aware launch operator for agent-built products. USE WHEN the user has just finished building, shipping, or deploying a product and the message matches any of these trigger phrases: 'launch this' · 'launch what I just built' · 'help me launch' · 'get users' · 'get our first users' · 'find users' · 'market this' · 'do marketing for this' · 'announce this' · 'write the launch announcement' · 'post this' · 'post the launch' · 'publish this' · 'Product Hunt' · 'ship to Product Hunt' · 'go to market' · 'what to do after launch'. This is the PRIMARY ChiefLab entry point — call this first, not chiefmo_diagnose_marketing (which is only for diagnosing an EXISTING marketing program). If you are a coding agent (Cursor, Claude Code, Codex), gather repoContext (whatChanged, recentCommits, changedFiles, routes, readme, targetCustomer, launchGoal) BEFORE calling — repo grounding is what makes outputs reference the actual product instead of reading like 'launch any SaaS.' Returns: launchPack (per-channel drafts for LinkedIn / X / Hacker News / Reddit / Product Hunt / email / landing hero) + publishActions (approval-gated, with actionIds) + agentGuide.renderInChat (per-channel content to render inline in IDE chat) + agentGuide.nextToolCalls.primary.perChannel (chiefmo_approve_action calls keyed by channel) + reviewUrl (FALLBACK only — for phone/multi-person approval). IDE-NATIVE FLOW: render each channel's draft inline in chat, wait for user to say 'approve ' or 'approve all', call chiefmo_approve_action per approved action. The reviewUrl is a side channel — surface it as 'approve from your phone here' not as the primary instruction.
| Name | Required | Description | Default |
|---|---|---|---|
| icp | No | Optional ICP block. Overrides repoContext.targetCustomer + brand.audience for ICP grounding. | |
| brand | No | OPTIONAL but recommended. Pass the brand context inline (name, audience, voice, pillars, competitors, visualCues). Cheapest path: $0, instant. ChiefLab uses this directly instead of scanning. See chiefmo_diagnose_marketing for the full brand schema. | |
| channels | No | Which launch channels to generate. Default: ["linkedin", "x", "product_hunt", "email", "landing_hero"]. Pass a subset to skip channels you don't want. | |
| tenantId | No | Optional. The end-user this launch is for. Loads stored brand context + voice samples + memory if you've called chieflab_set_tenant_context before. | |
| outputMode | No | "context" (default, ~$0): per-channel briefs your agent's LLM renders into final copy. "full": ChiefLab generates the actual posts server-side (premium credits, slower, useful when caller has no LLM). | |
| productUrl | No | The URL of the product being launched. Used as primary signal for positioning + brand extraction. Pass either this or productDescription (or both). | |
| templateId | No | Optional. Apply a saved launch template (chieflab_list_launch_templates → chieflab_save_launch_template). Template channels / launchAngle / brand / repoContext fill missing values; caller-provided values still win. | |
| webhookUrl | No | Optional HTTPS URL ChiefLab POSTs to when launch pack is ready, when actions are approved, and when posts go live. Signed with X-ChiefLab-Signature. | |
| launchAngle | No | Optional. If you already know the launch angle ("ship-friend launch", "product hunt blitz", "founder-led whisper"), pass it. Otherwise ChiefLab derives one from the product. | |
| productType | No | Optional override for product-type detection (devtool | saas | ai_agent | marketplace | internal_tool | consumer). Without this, ChiefLab auto-detects from URL + description + repoContext signals. | |
| repoContext | No | STRONGLY RECOMMENDED for agent-built products. Pass the codebase context for the product being launched — the more grounding ChiefLab has, the less generic the launch pack. Without this, outputs read like 'launch any SaaS.' With it, outputs reference the actual product. Coding agents (Cursor / Claude Code / Codex / Bolt / Lovable) should gather this from the user's repo before calling. | |
| scheduleFor | No | Optional ISO 8601 timestamp. If passed, publish actions schedule for this time instead of immediate-on-approval. UTC. | |
| imagesNeeded | No | How many on-brand launch graphics to generate via Gemini. Default: 1 (one hero graphic). 0 to skip image gen entirely. Image gen is metered separately (~$0.04 cost / ~$0.10 retail per image). | |
| approverEmail | No | Optional. The human who will approve publish actions. Surfaced in the onboarding checklist. | |
| idempotencyKey | No | Optional dedupe key. Critical for launches — prevents double-publish if the agent retries. | |
| productDescription | No | One-paragraph description of what the product is, who it's for, and why it matters. Use when productUrl isn't ready yet (pre-launch, private repo, internal tool). |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations, so description carries full burden. Describes return values (launchPack, publishActions, agentGuide), flow (render inline, wait for approval), and side channel (reviewUrl). Mentions idempotencyKey to prevent double-publish. Minor gap: does not explicitly state it's a read-only or mutation operation.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Well-structured with sections (triggers, primary entry, returns, IDE flow), but verbose. Multiple sentences could be tightened without losing meaning.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given complexity (16 params, nested objects, no output schema), description covers purpose, usage, behavioral traits, and parameter details comprehensively. The no output schema is compensated by describing return structure explicitly.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%, baseline 3. The description adds significant value beyond schema: explains purpose, trade-offs (outputMode vs context), pre-conditions (tenant context), and practical advice (cheapest path for brand).
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Clearly states 'Repo-aware launch operator for agent-built products' with specific trigger phrases. Distinguishes from sibling chiefmo_diagnose_marketing by stating it's the primary entry point and not for diagnosing existing programs.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicitly lists when to use (trigger phrases) and when not to (diagnose_marketing only for existing programs). Provides guidance for coding agents to gather repoContext before calling.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_list_actionsAInspect
List ChiefMO proposed actions, optionally filtered by run id.
| Name | Required | Description | Default |
|---|---|---|---|
| runId | No |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden but only states the basic purpose. It does not disclose behavioral traits such as pagination, ordering, authorization requirements, or the state of actions listed (e.g., pending, approved). This is insufficient for an agent to fully understand the tool's behavior.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single sentence, perfectly concise with no wasted words. It is front-loaded with the action and resource, and the filter condition is appended efficiently.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's simplicity (one optional param, no output schema), the description covers the basic functionality. However, it omits context about what 'proposed actions' are, the return format, or any behavioral details, leaving gaps for an agent to fully leverage the tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has one parameter 'runId' with 0% description coverage. The description adds that it is optional and used for filtering, which provides minimal additional meaning. However, it does not specify expected format or semantics of the filter, making it adequate but not enriched.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description clearly states the verb 'List' and resource 'ChiefMO proposed actions', with an optional filter by run id. Among siblings like 'chiefmo_get_action' and 'chiefmo_list_runs', it is distinct in listing proposed actions.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage (listing actions, optionally filtered), but provides no explicit guidance on when to use this tool versus alternatives like 'chiefmo_get_action' for a single action or 'chiefmo_list_runs' for runs. No when-not or alternatives are mentioned.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_list_assetsAInspect
List ChiefMO generated assets, optionally filtered by run id.
| Name | Required | Description | Default |
|---|---|---|---|
| runId | No |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided. Description indicates listing behavior and mentions filtering but omits details like pagination, permissions, or side effects. It is adequate for a simple read operation.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Single sentence, no unnecessary words, front-loaded with action and resource. Perfectly concise.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no output schema and one parameter, the description is complete enough for a list operation. However, it lacks definition of 'assets' and return format, slightly reducing completeness.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Single parameter 'runId' has 0% schema description coverage. The description clarifies it as an optional filter, which adds meaningful context beyond the schema's type-only definition.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description clearly states the tool lists ChiefMO generated assets with optional filtering by runId. It distinguishes from sibling 'chiefmo_get_asset' which likely retrieves a single asset.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Description mentions optional filtering by runId, guiding when to use that parameter. It does not explicitly exclude cases or mention alternatives, but the purpose is well implied.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_list_runsBInspect
List recent ChiefMO runs.
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, and the description does not disclose behavioral traits such as pagination, ordering, or the meaning of 'recent'. It assumes general knowledge about listing operations.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is very concise at three words, but it is not under-specified for a simple list tool with no parameters. A bit more detail could enhance clarity without adding bulk.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the absence of an output schema, the description should explain the return format, sorting, and recency criteria. It fails to provide this context, leaving the agent with incomplete information.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
With zero parameters, the baseline is 4. The description states 'recent', implying a default time filter. This adds minimal meaning beyond the empty schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states it lists recent ChiefMO runs, distinguishing it from sibling tools like chiefmo_get_run which retrieves a single run. However, 'recent' is vague and could be more specific.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
No guidance on when to use this tool versus alternatives like chiefmo_get_run or other list tools. The description lacks context on prerequisites or exclusions.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_list_skillsAInspect
List internal ChiefMO marketing skills available to the agent.
| Name | Required | Description | Default |
|---|---|---|---|
No parameters | |||
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, and the description offers no behavioral details beyond the basic function. It doesn't state that the tool is read-only, whether it can return an empty list, or any other traits. The agent gains no insight into side effects or constraints.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single sentence with no extraneous words. It is front-loaded and efficient, conveying the essential purpose without any wasted content.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a parameterless tool with no output schema or annotations, the description is minimal but adequate. It covers the core function, but lacks context about the format or scope of the returned skills, which could be important for an agent deciding how to use the result.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has no parameters, so schema description coverage is 100%. The description adds no parameter-level detail, but none is needed. It correctly implies the tool requires no input.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description uses a specific verb ('List') and identifies the resource as 'internal ChiefMO marketing skills'. It clearly distinguishes from sibling tools which are about launching, connecting, or other actions, making the tool's purpose unambiguous.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
There is no guidance on when to use this tool versus alternatives. With many sibling tools, explicit context such as 'Use this to discover available skills before executing marketing tasks' would help, but it's absent.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_magic_editAInspect
USE WHEN the user wants a quick tweak to existing marketing copy — 'shorter', 'warmer tone', 'change CTA to question', 'more direct', 'remove the third bullet', 'make it founder-led not corporate'. Apply a one-shot edit to a piece of copy. Default outputMode 'context' returns a small edit-brief your agent's LLM applies (~$0 token cost). Pass outputMode 'full' to have ChiefLab re-render server-side (premium credits). Use for rapid iteration without re-running the full diagnose flow.
| Name | Required | Description | Default |
|---|---|---|---|
| brand | No | Optional inline brand context (same shape as chiefmo_diagnose_marketing.brand). Cheaper than tenantId lookup if your agent already has the data. | |
| original | Yes | The text to edit. Whatever your agent's LLM produced last time, paste it here. | |
| tenantId | No | Optional. If passed, ChiefLab applies the tenant's brand voice + voice samples + memory to the edit instruction. | |
| outputMode | No | "context" (default, ~$0): returns an edit-brief your LLM applies. "full": ChiefLab re-renders server-side (premium credits, slower). | |
| contentType | No | Optional. What the original is (linkedin_post, x_post, email, landing_hero, ad_copy, etc.). Helps the brief target platform-specific constraints. | |
| instruction | Yes | What to change. Examples: "make it shorter", "warmer tone", "change CTA to a question", "remove the third bullet", "more founder-led, less corporate". |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden of behavioral disclosure. It explains the outputMode behavior ('context' returns an edit-brief, 'full' re-renders server-side), cost implications (~$0 token cost vs premium credits), and the quick iteration nature. It does not mention authorization or idempotency, but the non-destructive nature of editing copy is implied.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single paragraph with front-loaded usage guidance and examples. Every sentence adds value, covering purpose, usage, parameters, and cost. No redundant or vague language, achieving high density of useful information.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool complexity (6 params, nested objects, no output schema or annotations), the description adequately explains what the tool does, its inputs, and its output modes. It mentions the edit-brief or full re-render as outputs, but does not detail the exact format of the edit-brief (e.g., whether it returns a JSON structure or just text). This is a minor gap.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%, so baseline 3. The description adds significant value beyond the schema: it explains the 'brand' parameter as a cheaper alternative to tenantId, details the 'outputMode' choices and their cost/performance trade-offs, provides examples for 'instruction', and clarifies that 'original' should be the previous LLM output. This greatly aids parameter selection.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states 'USE WHEN the user wants a quick tweak to existing marketing copy' and gives concrete examples like 'shorter', 'warmer tone'. It uses specific verb+resource ('apply a one-shot edit to a piece of copy') and distinguishes from sibling tools by noting it avoids 're-running the full diagnose flow'.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description explicitly says 'USE WHEN' and lists usage examples. It also references the alternative diagnose flow, providing context for when to use this tool instead. However, it does not explicitly state when NOT to use or list prerequisites like needing a previous diagnose, which would strengthen the guidance.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_measure_launch_resultsAInspect
Pull post-launch metrics + recommend next iteration. USE WHEN it's been ≥24h since chiefmo_publish_approved_post fired and the user asks 'how did the launch perform' or 'what should we do next.' Returns Zernio engagement (likes / comments / views / shares per account) + GA4 traffic if connected + Search Console queries if connected + a structured next-move recommendation brief your LLM renders into a customer-facing summary. The closing tool in the closed loop: build → launch → publish → measure → iterate.
| Name | Required | Description | Default |
|---|---|---|---|
| runId | Yes | The launchId returned by chiefmo_launch_product. Required. | |
| lookbackDays | No | How many days of history to consider (default 7, max 90). |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description carries full burden. It describes outputs (metrics, recommendation brief) and notes the recommendation is a brief rendered into summary. However, it doesn't explicitly state read-only nature or side effects, which are implied but not confirmed.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Single paragraph that front-loads purpose and usage. Comprehensive but could be slightly trimmed without losing meaning. Every sentence adds value.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given 2 parameters, high schema coverage, and no output schema, description adequately explains return values (metrics + recommendation brief) and positions tool within a closed loop. No missing critical information for agent decision.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100% with descriptions. Description adds context by linking runId to chiefmo_launch_product's return value, aiding agent understanding beyond schema. No additional details for lookbackDays, but schema already sufficient.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description clearly states 'Pull post-launch metrics + recommend next iteration' and lists specific metrics (Zernio engagement, GA4, Search Console). It distinguishes itself as 'the closing tool in the closed loop', setting it apart from sibling tools.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicitly states when to use: 'USE WHEN it's been ≥24h since chiefmo_publish_approved_post fired and the user asks how did the launch perform or what should we do next.' Provides clear trigger conditions.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_memory_diffAInspect
Report what the company brain has learned in a window. New repo facts, proof points, channel performance, persona/competitor updates. Use as a weekly review or to show stakeholders why the moat compounds.
| Name | Required | Description | Default |
|---|---|---|---|
| goal | Yes | Window description (e.g. 'last 14 days', 'since launch v2'). | |
| brandUrl | No | Optional brand URL. | |
| sinceIso | No | Optional ISO timestamp; defaults to 30 days ago. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries full burden. It implies a read-only report but does not explicitly state behavioral traits such as being non-destructive or requiring specific permissions. The description is adequate but could be more transparent about side effects.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is two sentences, front-loads the main purpose, lists categories, and provides usage context. Every sentence is valuable with no redundancy.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the absence of an output schema, the description gives a reasonable sense of what the report contains (categories of learning) but could be more specific about the return format or structure. Still, it is largely sufficient for an AI agent to understand the output.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, and the tool description does not add significant new meaning beyond the parameter descriptions. For example, the 'goal' parameter is already well described in the schema with examples.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool reports what the company brain has learned over a window, listing specific categories (repo facts, proof points, etc.) and use cases (weekly review, stakeholder communication). It distinguishes from sibling tools like chiefmo_company_brain by focusing on temporal changes.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description explicitly suggests use cases ('Use as a weekly review or to show stakeholders'), providing clear context for when to employ the tool. However, it does not mention alternatives or when not to use it, which would strengthen guidance.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_monitor_competitorsAInspect
Snapshot competitor positioning, recent content, and pricing signals. Returns response-angles per competitor and white-space opportunities for the brand. Re-run monthly; deltas surface in chieflab_competitor_watch.
| Name | Required | Description | Default |
|---|---|---|---|
| goal | Yes | Which competitors to track and what to watch for. | |
| brandUrl | No | Optional brand URL. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It states what the tool does and returns but does not explicitly declare it as read-only or outline any side effects. While the output description suggests a non-destructive snapshot, behavioral traits like mutability are not addressed.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is three sentences long, front-loaded with the core action, and every sentence adds value. No redundant or irrelevant information.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no output schema, the description adequately explains return values (response-angles, white-space opportunities). It provides usage cadence and downstream context for deltas, making it self-contained for a monitoring tool.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100% with descriptions for both parameters, meeting the baseline. The description adds context about what the tool produces, which helps interpret the parameters, but does not extend the schema's meaning significantly.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly identifies the tool's function: snapshotting competitor positioning, recent content, and pricing signals, and returning response-angles and white-space opportunities. It distinguishes itself from sibling tools (mostly action-oriented marketing tools) as a monitoring/analysis tool.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description explicitly recommends monthly re-running and directs users to 'chieflab_competitor_watch' for deltas, providing clear guidance on usage cadence and where to find changes. It does not explicitly state when not to use, but the context implies periodic monitoring.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_plan_content_calendarAInspect
Build a 4-week multi-channel content calendar (channel cadence, deadlines, approval gates, re-plan triggers). Manual mode emits markdown tables; connected mode (Zernio/Resend) auto-fires approved items on their scheduled dates.
| Name | Required | Description | Default |
|---|---|---|---|
| goal | Yes | Calendar scope — e.g. 'Q2 launch month' or '4 weeks of pre-PMF founder content'. | |
| brandUrl | No | Optional brand URL. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations, the description fully explains behavior: manual vs connected mode, output format (markdown tables), auto-firing via Zernio/Resend, and mentions deadlines, approval gates, re-plan triggers. It lacks info on destructive actions or auth needs.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences, front-loaded with the core action, no wasted words. Each sentence adds distinct value: mode selection and output format.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no output schema, the description explains output format. It covers key aspects like schedule, approvals, triggers, and modes. Could mention if the calendar persists, but overall adequate.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%, so the description adds minimal value beyond the schema. The 'goal' parameter has an example but no syntax details. Baseline 3 is appropriate.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool builds a 4-week multi-channel content calendar. The verb 'plan' and resource 'content calendar' are specific and differentiate it from siblings, none of which are calendar planners.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description mentions manual and connected modes, implying when each might be used, but does not explicitly state when to use this tool versus alternatives or when not to use it.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_post_launch_reviewAInspect
USE WHEN ≥24h has passed since chiefmo_publish_approved_post fired and the user asks 'how did the launch perform?', 'what worked?', 'metrics from my launch'. Closes the loop: pulls 24h+ engagement (likes / comments / views / shares / follower delta) for posts in the launch run + recommends next iteration (rewrite, refresh creative, double down, pause). Returns per-account analytics + top posts + best-time-to-post + a structured next-move recommendation brief.
| Name | Required | Description | Default |
|---|---|---|---|
| runId | Yes | The launchId returned by chiefmo_launch_product. Used to scope analytics to posts from that launch. | |
| accountIds | No | Optional. Specific Zernio accountIds to pull analytics for. Defaults to all accounts that posted in this launch. | |
| outputMode | No | "context" (default): returns metrics + a brief your LLM renders into the customer-facing recommendation. "full": ChiefLab generates the recommendation server-side. | |
| lookbackDays | No | How many days of history to consider (default 7, max 90). |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description covers behavioral traits: it pulls 24h+ engagement data, recommends next iteration, and returns per-account analytics. It doesn't mention side effects or authentication needs, but the read-only nature is implied by 'closes the loop'.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single paragraph but packs essential information: usage condition, action, returns, and recommendations. It front-loads the trigger condition and is relatively focused, though slightly verbose.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool has 4 parameters and no output schema or annotations, the description is fairly complete. It explains when to use, what metrics are pulled, and the structure of the recommendation brief. Minor gap: doesn't detail the exact output format, but sufficient for selection.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
All parameters have schema descriptions (100% coverage), so baseline is 3. Description adds value by explaining runId as the launchId from chiefmo_launch_product and clarifying defaults for accountIds and outputMode, providing useful context beyond the schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description clearly states it performs post-launch review, pulling engagement metrics and recommending next moves. It specifies the trigger condition (≥24h after chiefmo_publish_approved_post) and differentiates from similar siblings like chiefmo_measure_launch_results by focusing on closing the loop and providing strategic recommendations.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicitly states when to use: after 24h have passed since publish and user asks about launch performance. Provides clear context but doesn't explicitly state when not to use or mention alternatives, though the condition is well-defined.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_prepare_next_moveAInspect
USE WHEN the user has just published a launch (or ≥24h post-launch) and asks 'what's next?', 'follow-up post?', 'iterate on this', 'plan day 2'. P13 — turns one launch into a 7-day launch mode: drafts a queue of follow-up actions (day 2 reply post, day 3 case study, day 5 metrics share, day 7 retrospective) grounded in the prior launch's runId, the per-tenant brand voice + repo facts + channel performance from the P9 company brain, and the prior recommendation (so the next move doesn't repeat the last one). Returns a draft queue + signed reviewUrl + the standard agentGuide block. Each next-move action lands in awaiting_approval; nothing fires without explicit human approval.
| Name | Required | Description | Default |
|---|---|---|---|
| runId | Yes | The prior launchId returned by chiefmo_launch_product. The prepared queue grounds in this run's measurement + memory. | |
| horizon | No | "day_2" | "day_3" | "week" — how far to plan. Default: "week" (7-day mode). | |
| channels | No | Optional. Restrict next-moves to these channels. Defaults to the channels of the prior launch. | |
| workspaceId | No | Optional workspace id. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations provided, the description carries the full burden. It reveals that the tool drafts a queue that lands in 'awaiting_approval' and requires explicit human approval before any action fires. It also explains the grounding data (prior launch's runId, brand voice, repo facts) and the output format (draft queue, signed reviewUrl, agentGuide block). This is thorough for a mutation tool.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single paragraph that front-loads the usage condition. Every sentence provides useful information without redundancy. It is dense but clear, earning its place. Slight improvement could be made with explicit sections, but it is effective.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool has 4 parameters, no output schema, and no annotations, the description covers inputs, behavior, safety (human approval), and output (draft queue + reviewUrl + agentGuide). It lacks error handling details but is otherwise complete for an agent to understand and invoke correctly.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
All four parameters have descriptions in the input schema (100% coverage), so baseline is 3. The description adds value by explaining how runId is used as the prior launchId and that horizon defaults to 'week', but it does not provide additional syntax or format details beyond the schema. The added context is modest.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description starts with 'USE WHEN' and clearly states the tool's purpose: turning one launch into a 7-day launch mode by drafting a queue of follow-up actions. It specifies the trigger condition (launch published and user asks 'what's next?') and explains the output, distinguishing it from sibling tools like chiefmo_launch_product or chiefmo_continue_launch_loop.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description explicitly defines when to use this tool ('just published a launch' or '>=24h post-launch' and asking for next steps) and provides example user prompts. It does not explicitly state when not to use it or list alternatives, but the context is strong enough to guide appropriate usage.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_prepare_weekly_reportAInspect
Compose a weekly marketing report with KPI deltas, what changed and why, wins, risks, and recommended next 3 actions. Grounded in connector snapshots when available.
| Name | Required | Description | Default |
|---|---|---|---|
| goal | Yes | Reporting period and focus (e.g. 'last week's marketing across all channels'). | |
| brandUrl | No | Optional brand URL for context. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Without annotations, the description carries full burden. It describes output content but omits side effects, auth needs, execution behavior (e.g., real-time vs async), or whether it is read-only. Partially transparent but incomplete.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences, front-loading key report components and data source. No unnecessary words; efficient and clear.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a simple 2-param tool without output schema, the description adequately covers report content and data grounding. Missing details on prerequisites (e.g., connector availability) and output format, but overall sufficient.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100% with both parameters already described. The tool description adds no new parameter semantics beyond the schema, so baseline 3 applies.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool composes a weekly marketing report with specific components (KPI deltas, changes, wins, risks, next actions), distinguishing it from siblings like chiefmo_diagnose_marketing or chiefmo_synthesize_marketing_analytics.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies use for weekly reporting and mentions grounding in connector snapshots, but lacks explicit when-to-use or when-not-to-use guidance compared to alternatives. No exclusion criteria or prerequisites are stated.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_publish_approved_postAInspect
Publish an approved social post (LinkedIn / X / Threads / Instagram / Facebook / Bluesky / TikTok) via Zernio. USE WHEN the user has approved a publishAction from chiefmo_launch_product on the reviewUrl and you need to fire the post. Strict approval gate: returns { reason: 'requires_approval', reviewUrl } if the action isn't approved yet. Once executed, returns the live platform post URL. Money/external-system action — never call this without an approved actionId.
| Name | Required | Description | Default |
|---|---|---|---|
| content | Yes | The final rendered post text (your LLM rendered this from the brief in the launch pack). | |
| actionId | Yes | The publishAction.id from chiefmo_launch_product. Required. | |
| mediaUrls | No | Optional. Image/video URLs (or base64 data URLs from launch graphics). | |
| platforms | Yes | Which platforms + Zernio accountIds to publish to. Get accountIds from chieflab_list_publish_accounts. | |
| scheduleAt | No | Optional ISO 8601. Schedule for later instead of immediate. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description carries full burden. Discloses it's a money/external-system action, approval gate behavior, and output (live URL). Could mention idempotency or rate limits but still strong.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Very concise: two main sentences plus a punchy warning. Front-loaded with purpose, then usage, then behavior. No wasted words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
No output schema, but description explains both possible returns (requires_approval or live URL). Covers approval gate, platforms, scheduling, and links to related tools. Complete for a publish action.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%, but description adds value by explaining 'content' as final rendered text from brief, and 'platforms' references sibling tool for accountIds. Adds meaningful context beyond schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Clearly states verb 'publish' and resource 'approved social post' with platforms listed. Differentiates from siblings by requiring an approved actionId from chiefmo_launch_product.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicitly tells when to use (after user approves publishAction) and warns against calling without approval. Describes the strict approval gate and alternative return behavior.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_record_brand_voiceAInspect
Update the brand-voice facet of the company brain: banned phrases (never use), preferred claims (always available), tone descriptors, voice examples. Run after rejecting a draft or after a founder says 'don't write it like that.' Future runs inherit these constraints automatically.
| Name | Required | Description | Default |
|---|---|---|---|
| tenantId | No | ||
| visualStyle | No | ||
| workspaceId | No | ||
| bannedPhrases | No | Lowercase phrases future drafts must never contain. | |
| voiceExamples | No | ||
| preferredClaims | No | Specific claim phrasings this brand prefers. | |
| toneDescriptors | No |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full burden. It states it updates the brand-voice facet and that future runs inherit constraints. However, it does not disclose whether the update is destructive (overwrites vs merges), what happens to existing values, or if permissions are required. More detail on merge/replace behavior would improve transparency.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is two sentences. The first sentence lists the elements, the second gives usage triggers. It is concise, front-loaded, and every sentence adds value.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the complexity (7 params, nested objects, no output schema), the description is incomplete. It does not describe return values (void or something), nor does it explain the 'visualStyle' parameter present in schema. It also omits that no parameters are required. For a tool with such a rich schema, the description should cover more aspects.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema has 7 parameters with only 29% description coverage. The description lists the elements (banned phrases, preferred claims, tone descriptors, voice examples) but does not explain the nested objects like 'visualStyle' or the structure of 'voiceExamples'. Two of seven parameters are clarified by the description, but it does not fully compensate for the low schema coverage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool updates the brand-voice facet of the company brain, listing specific elements: banned phrases, preferred claims, tone descriptors, voice examples. It also provides example triggers (after rejecting a draft or founder feedback), which distinguishes it from siblings like 'chiefmo_record_voice_sample' that likely record individual samples.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description explicitly says when to run: 'Run after rejecting a draft or after a founder says "don't write it like that."' It also states that future runs inherit these constraints, implying it sets defaults. While it doesn't explicitly mention when NOT to use or name alternatives, the guidance is clear enough for an agent.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_record_channel_resultBInspect
Record one channel-performance row (channel × hook × format × CTA × outcome × metric). Powers the 'what's worked / hasn't' block in future plans. Outcome is one of: won | lost | inconclusive | killed.
| Name | Required | Description | Default |
|---|---|---|---|
| cta | No | ||
| hook | No | ||
| notes | No | ||
| format | No | post | thread | carousel | video | image | long_form | newsletter | |
| channel | Yes | linkedin | x | email | blog | ad | reddit | threads | instagram | product_hunt | |
| outcome | No | won | lost | inconclusive | killed (default: inconclusive) | |
| baseline | No | ||
| tenantId | No | ||
| sourceRunId | No | ||
| workspaceId | No | ||
| primaryValue | No | ||
| primaryMetric | No | ||
| sourceAssetId | No | ||
| sourceActionId | No |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided; description carries full burden. States it records a row (write operation) but lacks details on side effects, permissions, rate limits, or what happens to existing data.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences, no redundant information. First sentence captures core purpose, second adds outcome options. Efficient and front-loaded.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
High complexity (14 params, no output schema, low schema coverage) but description is too brief. Lacks crucial context about parameter meanings, return values, and usage patterns.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is low (21%). Description adds context on dimensions but fails to explain many parameters (e.g., notes, baseline, primaryMetric, source IDs). Parameter semantics remain unclear.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Description clearly states it records one channel-performance row with dimensions (channel × hook × format × CTA × outcome × metric) and explains its use in future plans. Distinguishes from sibling recording tools like chiefmo_record_brand_voice.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Provides context for use (powers 'what's worked / hasn't' block) and lists allowed outcome values. However, no explicit when-to-use vs alternatives or when-not-to-use guidance.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_record_personaAInspect
Capture an audience/persona memory entry: ICP, segments, top objections, proof points, banned claims, winning angles. Future runs ground against this memory. Use early in onboarding, then refresh when your understanding shifts.
| Name | Required | Description | Default |
|---|---|---|---|
| goal | Yes | Persona context — what you've learned about who buys / signs up. | |
| brandUrl | No | Optional brand URL. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
Discloses that future runs ground against this memory, indicating persistence and relevance. With no annotations, the description should also cover side effects like overwriting or conflicts, but it lacks those details. Not harmful but incomplete.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences, front-loaded with the action and content, then usage guidance. No unnecessary words.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Covers purpose and usage timing but lacks details on return values or persistence semantics. Given no output schema and no annotations, more context on behavior would improve completeness.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100%, so parameters are defined. The description adds examples (ICP, segments) but no format or constraints beyond the schema. Baseline 3 is appropriate.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states it captures an audience/persona memory entry with specific content like ICP and segments. It distinguishes from generic memory tools by naming persona-specific fields, though it doesn't explicitly differentiate from sibling memory tools like chiefmo_gtm_memory_set.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicitly advises use 'early in onboarding' and 'refresh when understanding shifts,' providing clear timing. However, it does not mention when not to use this tool or suggest alternatives among sibling memory tools.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_record_proofAInspect
Capture a structured proof asset (testimonial, metric, customer quote, case study, launch result, review, press) into the company brain. Future runs can cite it instead of inventing claims.
| Name | Required | Description | Default |
|---|---|---|---|
| body | Yes | Verbatim quote or 1-2 sentence summary. | |
| kind | Yes | testimonial | metric | customer_quote | case_study | launch_result | review | press | |
| title | No | ||
| tenantId | No | ||
| sourceUrl | No | ||
| metricName | No | Required when kind=metric or launch_result. | |
| metricUnit | No | users | USD | % | pct_pts | etc. | |
| observedAt | No | ISO date; defaults to now. | |
| attribution | No | Who/where (e.g. 'Jane Doe, CTO at Acme'; 'Reddit /r/SaaS'). | |
| metricValue | No | ||
| workspaceId | No |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries full behavioral burden. It indicates a write operation (capture) but does not disclose side effects, authentication needs, idempotency, or what happens on duplicate – leaving significant gaps.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two efficient sentences with no fluff, front-loading the action and purpose. Every word earns its place.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a tool with 11 parameters and no output schema, the description gives high-level purpose but lacks details on return values, parameter dependencies, or constraints. It is adequate but not comprehensive; agents may need to infer usage from the schema alone.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 55%. The description lists example kinds (testimonial, metric, etc.), aligning with the 'kind' parameter, but does not explain other parameters like title, sourceUrl, metricName, etc. It adds some semantic value but not enough to compensate for the low parameter coverage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool captures structured proof assets (testimonial, metric, etc.) into the company brain, with a specific purpose: 'Future runs can cite it instead of inventing claims.' This verb+resource purpose distinguishes it from sibling tools like chiefmo_check_claims or other record tools.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage when you have evidence to store for citation, contrasting with inventing claims. It provides context but does not explicitly state when not to use or mention direct alternatives among siblings, though the set of proof kinds suggests appropriate scenarios.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_record_repo_factAInspect
Persist one structured product fact (feature / route / pricing / integration / changelog / screenshot / metric / constraint) extracted from the repo. Older facts with the same (kind, name) are auto-superseded. Run on every launch so the brain stays current with the codebase.
| Name | Required | Description | Default |
|---|---|---|---|
| kind | Yes | feature | route | pricing | integration | changelog | screenshot | metric | constraint | |
| name | Yes | ||
| value | No | ||
| details | No | ||
| tenantId | No | ||
| sourceFile | No | ||
| workspaceId | No | ||
| sourceCommit | No | ||
| sourceRepoUrl | No |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations, the description must disclose behavior. It states 'Older facts with the same (kind, name) are auto-superseded', which is a key behavioral trait. However, it does not mention idempotency, return values, or failure handling, leaving gaps.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two sentences, no extraneous words, direct and front-loaded with the purpose. Every sentence adds value.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Despite the auto-supersede note, the description lacks context on prerequisites, output, data relationships, or integration with sibling tools like chiefmo_extract_repo_facts. With 9 parameters and no annotations, more detail is needed for completeness.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is only 11%, and the description lists the allowed kinds for 'kind' but that is already in the schema description. The description adds no additional meaning for other parameters like 'name', 'value', 'details', etc., failing to compensate for low coverage.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool persists a structured product fact, listing specific kinds (feature, route, etc.) and distinguishes it from extraction tools by mentioning 'extracted from the repo'. The verb 'persist' and resource 'structured product fact' are specific.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description says 'Run on every launch' which implies periodic use but does not explicitly state when to use this tool vs alternatives like chiefmo_record_proof. No when-not-to-use guidance is provided.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_reject_runCInspect
Reject a ChiefMO run that is awaiting approval.
| Name | Required | Description | Default |
|---|---|---|---|
| id | Yes |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations, the description must fully disclose behavioral traits. It mentions the prerequisite that the run must be awaiting approval, but it does not cover side effects, permissions required, reversibility, or what happens after rejection.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is a single, clear sentence with no redundant information. It is appropriately concise and front-loaded, making it efficient for quick parsing.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the tool's simplicity (one parameter, no output schema, no annotations), the description is incomplete. It lacks essential details about the rejection outcome, error scenarios, and how to obtain the run ID, which an AI agent would need for reliable invocation.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
The input schema has one required parameter 'id' with no description, and the tool description does not add any meaning about the parameter, such as what the ID represents or where to obtain it. Schema coverage is 0%, and the description fails to compensate.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the verb 'Reject' and the resource 'a ChiefMO run that is awaiting approval.' It effectively distinguishes this tool from siblings like 'chiefmo_approve_run' and 'chiefmo_get_run' by specifying the action and the condition of the run.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides no guidance on when to use this tool versus alternatives, nor does it mention prerequisites, constraints, or when not to use it. It only states what it does without contextual usage advice.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_render_quote_cardAInspect
Render a deterministic on-brand quote card as PNG using Era (HTML/CSS layout + Playwright render), not diffusion. USE WHEN the user asks for a quote card, founder quote graphic, testimonial graphic, text-heavy launch visual, or brand-precise social graphic. Takes brand context and exact quote text, returns a PNG dataUrl plus layout metadata. Local pipeline; no publish/send side effect.
| Name | Required | Description | Default |
|---|---|---|---|
| brand | Yes | Brand context. Supports { name, voice, palette: { background, foreground, accent }, typography: { heading, body }, markUrl } or legacy visualCues.palette/type. | |
| quote | Yes | Exact quote text to render. Era renders this verbatim. | |
| returnHtml | No | Whether to include the generated HTML in the response. Default false to keep MCP payloads smaller. | |
| attribution | Yes | Who said it, or a short source label. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
With no annotations, the description fully discloses the rendering pipeline (HTML/CSS + Playwright), determinism, local processing, and lack of side effects. It clearly explains what the tool does and does not do.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is very concise with no wasted words. It front-loads the key action and then efficiently covers usage, inputs, and outputs in a few sentences.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no output schema, the description adequately explains the return format (PNG dataUrl + layout metadata). It also covers the tool's scope and constraints, making it complete for an agent to invoke correctly.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%, baseline 3. The description adds value by explaining the verbatim rendering of quotes, the brand object structure with examples, and the default for returnHtml. This enhances understanding beyond the schema.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool renders a deterministic quote card as PNG using Era, not diffusion. It lists specific use cases (quote card, founder quote, testimonial, etc.) and distinguishes from other methods, making its purpose unmistakable.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description provides explicit 'USE WHEN' conditions and states that it has no publish/send side effect, guiding the agent to choose this tool for text-heavy brand graphics and avoid it for publishing actions.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_repurpose_contentAInspect
Take a single source asset (launch post, blog, demo, customer proof) and emit platform-tuned variants for LinkedIn, X thread, Threads, and email. Each variant leads with a different hook — not paraphrases. Manual mode pastes; connected mode (Zernio/Resend) can publish on a cadence after approval.
| Name | Required | Description | Default |
|---|---|---|---|
| goal | Yes | Describe the source content and what you want repurposed. | |
| brandUrl | No | Optional brand URL. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It discloses that it generates variants with different hooks and can publish on approval, but it does not explicitly state that it is a read-only operation or what happens to the source asset. It is adequate but could be more explicit about safety and side effects.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Three sentences, front-loaded with the core purpose, then specific details about hooks and modes. No unnecessary words; each sentence earns its place.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given no output schema, the description explains the output as 'platform-tuned variants' but does not specify the format (e.g., text blobs, structured data). For a simple tool, it is mostly complete but could clarify return structure.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100%, and the description aligns with the schema by mentioning the goal parameter context. It adds value by describing the output (variants with hooks) and the two modes, which goes beyond the schema alone.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states it takes a single source asset and produces platform-tuned variants for specific platforms (LinkedIn, X, Threads, email) with unique hooks, which distinguishes it from sibling tools like chiefmo_generate_social_posts.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
It explains two modes (manual paste vs. connected publish with approval), providing context on when to use each. However, it lacks explicit exclusions or alternatives, though sibling differentiation is implied.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_send_approved_emailAInspect
Send an approved launch email via Resend. USE WHEN the user has approved an email publishAction from chiefmo_launch_product and you need to fire the send. Strict approval gate (same shape as chiefmo_publish_approved_post). from MUST be on a verified Resend domain — check chieflab_list_email_senders first. Single recipient or short list (≤50). Money/external-system action — once sent cannot be unsent.
| Name | Required | Description | Default |
|---|---|---|---|
| to | Yes | Recipient(s). Single email or array (≤50). | |
| from | Yes | Sender. Must be a verified Resend domain address (e.g. 'Brand <hi@yourdomain.com>'). | |
| html | No | HTML body (preferred for inboxing). | |
| text | No | Plain-text fallback (recommended). | |
| replyTo | No | Optional reply-to address(es). | |
| subject | Yes | ||
| actionId | No | The publishAction.id from chiefmo_launch_product. Required for the approval gate to fire. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations provided, so description carries full burden. Discloses that this is a 'money/external-system action,' that it cannot be unsent, and requires an approval gate. Also notes domain verification requirement.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Four sentences, no extraneous words. Front-loaded with core purpose and usage conditions. Each sentence adds unique, non-redundant value.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
No output schema, but as a side-effect action (send email), description fully covers needed context: approval gate, domain verification, constraints, and irreversibility. Complete for agent decision-making.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Adds significant context beyond input schema: explains that `actionId` is required for the approval gate, `from` must be verified, `to` can be single or array ≤50, and `html` is preferred while `text` is recommended. Schema coverage is high (86%), but description adds crucial behavioral meaning.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
Clearly states it sends an approved launch email via Resend, with specific verb and resource. Distinguishes from sibling tools by referencing similarity to chiefmo_publish_approved_post and mentioning the approval gate from chiefmo_launch_product.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
Explicitly instructs when to use: 'when the user has approved an email publishAction from chiefmo_launch_product.' Provides prerequisites (check verified domain via chieflab_list_email_senders) and constraints (≤50 recipients, irreversible action).
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_seo_briefAInspect
Generate a Search Console-grounded SEO refresh brief: which pages to refresh, which queries to target, content clusters to build, internal-link opportunities. Manual mode emits a structural brief; with Search Console connected, every recommendation cites actual impression / CTR / position numbers.
| Name | Required | Description | Default |
|---|---|---|---|
| goal | Yes | SEO question or target (e.g. 'recover ranking on /pricing', 'plan next quarter's content clusters'). | |
| brandUrl | No | Optional brand URL. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the burden. It discloses that the tool generates a brief and, when Search Console is connected, cites actual metrics. It does not mention destructive actions or side effects, but the tool appears to be a read/generation operation. The description is sufficient but could be more explicit about privacy or data handling.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
The description is two sentences, front-loaded with the core purpose, and each part adds value. No wasted words. It is appropriately sized for the tool's complexity.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
Given the moderate complexity (2 params, no output schema), the description covers the tool's purpose, modes, and output components. It does not specify the exact format or length of the brief, but it is sufficient for an AI agent to understand when to use it. Slightly more detail on the output structure would improve completeness.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema description coverage is 100% (both 'goal' and 'brandUrl' have descriptions). The description adds context that 'goal' is an SEO question or target, but it does not significantly extend beyond the schema. Baseline 3 is appropriate.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states it generates an SEO refresh brief with specific outputs (pages to refresh, queries to target, content clusters, internal-link opportunities) and distinguishes two modes (manual vs Search Console-connected). This is a specific verb-resource pair and differentiates from sibling tools like 'chiefmo_draft_blog_post' or 'chiefmo_plan_content_calendar'.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies when to use the tool (for SEO refresh briefs) and outlines two operational modes, but does not explicitly state when not to use it or provide alternative tool names. The context is clear, but exclusions are missing.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
chiefmo_synthesize_marketing_analyticsAInspect
Cross-channel marketing analytics synthesis. Pulls social / email / blog / GA4 / Search Console signals into a single 'what worked, what didn't, what to do next' story. Ends with 3 prioritized actions and memory updates to feed forward.
| Name | Required | Description | Default |
|---|---|---|---|
| goal | Yes | Reporting period or focus (e.g. 'last 30 days across all channels'). | |
| brandUrl | No | Optional brand URL. |
Tool Definition Quality
Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?
No annotations are provided, so the description carries the full burden. It indicates the tool synthesizes data and ends with prioritized actions and memory updates, suggesting both reading and writing. However, it does not disclose potential side effects (e.g., destructive effects), failure modes, or safety considerations beyond memory updates.
Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.
Is the description appropriately sized, front-loaded, and free of redundancy?
Two concise sentences cover the main action, data sources, output (story with actions), and memory updates. No wasted words, front-loaded with key verb and resource.
Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.
Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?
For a tool with two simple parameters and no output schema, the description covers core functionality adequately. However, it lacks details on the output format (beyond 'story') and how memory updates work, which are minor gaps given no annotations.
Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.
Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?
Schema coverage is 100% with both parameters described. The description adds context by listing data channels, but this does not significantly enhance parameter meaning beyond what the schema provides. Baseline of 3 is appropriate.
Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.
Does the description clearly state what the tool does and how it differs from similar tools?
The description clearly states the tool's function: cross-channel marketing analytics synthesis, producing a narrative with actions and memory updates. It identifies specific data sources (social, email, blog, GA4, Search Console), which distinguishes it from sibling tools like 'chiefmo_diagnose_marketing', though it does not explicitly contrast them.
Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.
Does the description explain when to use this tool, when not to, or what alternatives exist?
The description implies usage when cross-channel synthesis is needed by listing data sources, but it does not provide explicit guidance on when not to use this tool or suggest alternatives. The context is clear but lacks exclusions.
Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.
Claim this connector by publishing a /.well-known/glama.json file on your server's domain with the following structure:
{
"$schema": "https://glama.ai/mcp/schemas/connector.json",
"maintainers": [{ "email": "your-email@example.com" }]
}The email address must match the email associated with your Glama account. Once published, Glama will automatically detect and verify the file within a few minutes.
Control your server's listing on Glama, including description and metadata
Access analytics and receive server usage reports
Get monitoring and health status updates for your server
Feature your server to boost visibility and reach more users
For users:
Full audit trail – every tool call is logged with inputs and outputs for compliance and debugging
Granular tool control – enable or disable individual tools per connector to limit what your AI agents can do
Centralized credential management – store and rotate API keys and OAuth tokens in one place
Change alerts – get notified when a connector changes its schema, adds or removes tools, or updates tool definitions, so nothing breaks silently
For server owners:
Proven adoption – public usage metrics on your listing show real-world traction and build trust with prospective users
Tool-level analytics – see which tools are being used most, helping you prioritize development and documentation
Direct user feedback – users can report issues and suggest improvements through the listing, giving you a channel you would not have otherwise
The connector status is unhealthy when Glama is unable to successfully connect to the server. This can happen for several reasons:
The server is experiencing an outage
The URL of the server is wrong
Credentials required to access the server are missing or invalid
If you are the owner of this MCP connector and would like to make modifications to the listing, including providing test credentials for accessing the server, please contact support@glama.ai.
Discussions
No comments yet. Be the first to start the discussion!