Skip to main content
Glama

ChiefLab

Server Details

Open execution contract for agents doing business work. 6 operators, one MIT-licensed spec.

Status
Unhealthy
Last Tested
Transport
Streamable HTTP
URL

Glama MCP Gateway

Connect through Glama MCP Gateway for full control over tool access and complete visibility into every call.

MCP client
Glama
MCP server

Full call logging

Every tool call is logged with complete inputs and outputs, so you can debug issues and audit what your agents are doing.

Tool access control

Enable or disable individual tools per connector, so you decide what your agents can and cannot do.

Managed credentials

Glama handles OAuth flows, token storage, and automatic rotation, so credentials never expire on your clients.

Usage analytics

See which tools your agents call, how often, and when, so you can understand usage patterns and catch anomalies.

100% free. Your data is private.
Tool DescriptionsB

Average 4/5 across 91 of 91 tools scored. Lowest: 2.6/5.

Server CoherenceC
Disambiguation2/5

Many tools have overlapping purposes (e.g., multiple launch tools with aliases, multiple record tools). The existence of both chieflab_ and chiefmo_ prefixes for similar functions adds confusion. Descriptions help but are insufficient given the volume and redundancy.

Naming Consistency2/5

Inconsistent prefix usage (chieflab_ vs chiefmo_) with many aliases. While individual names follow verb_noun pattern, the mix of prefixes and duplicate functions (e.g., chieflab_brain_read and chiefmo_company_brain) breaks predictability.

Tool Count1/5

91 tools is far beyond the recommended range. The server attempts to cover a wide domain but could be streamlined significantly without losing functionality.

Completeness4/5

The tool set covers an impressively broad domain from signup to measurement, including connectors, publishing, analytics, and memory. Minor gaps exist (e.g., no direct tool for deleting a tenant or managing team roles beyond members) but overall it is comprehensive.

Available Tools

91 tools
chieflab_brain_readBInspect

Operator: chieflab-brain (primary read). Returns the per-tenant company brain: brand voice samples, repo facts, proof assets, channel performance memory, claim risk log. Drives cross-launch grounding. Spec-renamed alias of chiefmo_company_brain — both names route to the same handler. Record tools: chiefmo_record_proof / chiefmo_record_brand_voice / chiefmo_record_repo_fact / chiefmo_record_channel_result.

ParametersJSON Schema
NameRequiredDescriptionDefault
tenantIdNo
workspaceIdNo
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description must carry the full burden. It uses the word 'read' implying idempotency, but does not disclose side effects, authentication needs, rate limits, or input validation behavior. The description adds minimal transparency beyond the name itself.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is relatively concise, front-loading the purpose and providing useful alias and record tool context. While the list of record tools adds length, it is relevant for completeness. Slightly too verbose, but overall well-structured.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a simple read tool with no output schema, the description lists the components of the returned brain, which provides some context. However, it does not describe the output format or possible error states, leaving some gaps for an agent.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters1/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 0%, and the description does not explain the meaning or format of tenantId and workspaceId, nor how they affect the result. The only hint is 'per-tenant company brain,' but this does not map to the parameters explicitly.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool is a 'primary read' that returns the 'per-tenant company brain' with specific components (brand voice samples, repo facts, etc.), effectively distinguishing it from siblings. It also notes it is a spec-renamed alias of chiefmo_company_brain, which helps avoid confusion with the sibling tool.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies the tool is for reading contextual information before launches ('drives cross-launch grounding') and lists related record tools, but it does not explicitly guide when to use this tool versus alternatives or state when not to use it.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_connect_connectorAInspect

USE WHEN the user wants to connect a third-party data source via OAuth (HubSpot, Stripe, Intercom, Linear, etc.). Returns an authorize URL — surface to user → they grant access → connector becomes live for the workspace and ChiefLab reads real data instead of mock snapshots. NOTE: only chieflab_connect_publish_account (Zernio for social posting) is fully wired in v1; other connectors are scaffolded but the live-data path may be partial.

ParametersJSON Schema
NameRequiredDescriptionDefault
providerYesConnector provider id (hubspot, stripe, intercom, linear, etc.).
workspaceIdNoOptional workspace id (defaults to your API key's workspace).
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description carries the full burden. It clearly states the return value (authorize URL) and describes the OAuth flow and resulting connector state. It also warns about the partial live-data path for most connectors, providing useful behavioral context beyond the basic action.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single paragraph but is front-loaded with the key use case and includes necessary caveats. It is clear and direct, with no unnecessary words, though it could be slightly more structured.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

The description covers the tool's purpose, output (authorize URL), and limitations. Without an output schema, it adequately describes the return value. The note about v1 wiring adds important context for agent decision-making.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100%, so baseline is 3. The description does not add additional semantics beyond what the schema already provides for 'provider' and 'workspaceId'. It doesn't explain allowed values or usage details beyond the schema descriptions.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description starts with 'USE WHEN the user wants to connect a third-party data source via OAuth' and lists specific examples, clearly stating the tool's purpose and resource. It distinguishes from siblings by noting that only chieflab_connect_publish_account is fully wired.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines5/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description explicitly says 'USE WHEN' and provides context for connecting third-party data sources. It also includes a note about v1 limitations, guiding the agent on when this tool is appropriate and that other connectors may have partial functionality.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_connector_setup_stepsAInspect

Get step-by-step JIT setup instructions for a connector (zernio | resend | ga4 | search_console | github | cms). Returns prerequisites + ordered steps + verification commands. Use when a launch pack reports a connectorBlocker — pipe these steps to the user.

ParametersJSON Schema
NameRequiredDescriptionDefault
providerYeszernio | resend | ga4 | search_console | github | cms
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description carries full burden. It clearly indicates a read-only operation ('Get ... instructions') and describes the return value (prerequisites, steps, verification commands). No contradictions or missing behavioral details are apparent, though it could explicitly state no side effects.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two sentences: first covers purpose and output, second covers usage. No fluff, front-loaded with essential information. Every sentence earns its place.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness5/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a simple tool with one parameter and no output schema, the description fully covers purpose, input, output, and usage context. No gaps remain given the tool's complexity.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The schema has 100% coverage with a description listing valid providers. The description repeats these providers in parentheses but adds no additional semantic information beyond what the schema provides. Baseline of 3 is appropriate.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool's purpose: 'Get step-by-step JIT setup instructions for a connector (zernio | resend | ga4 | search_console | github | cms).' It specifies the exact output format (prerequisites, ordered steps, verification commands) and distinguishes itself from sibling connector tools like chieflab_connector_status by focusing on setup instructions.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description explicitly states when to use: 'Use when a launch pack reports a connectorBlocker — pipe these steps to the user.' This provides clear context and a specific trigger. It implies when not to use (other scenarios) but doesn't mention alternative tools, which is acceptable given the narrow scope.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_connector_statusAInspect

USE WHEN the user just completed a connector OAuth flow and you want to confirm it succeeded — 'did my connection work?', 'is my Zernio/HubSpot/Stripe connected?'. Returns status: pending | connected | failed for the given connectionId.

ParametersJSON Schema
NameRequiredDescriptionDefault
workspaceIdNo
connectionIdYes
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description carries full burden. It states returns status values (pending/connected/failed) but doesn't elaborate on auth needs, rate limits, or that it's a read-only operation. Adequate but could be more explicit.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Single sentence with context, example queries, and key return values. No filler, perfectly front-loaded, and efficient.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's simplicity (2 params, no output schema), the description gives enough context for its purpose. Could mention workspaceId's role, but overall it's clear for a status-check tool.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters2/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema has two parameters (workspaceId optional, connectionId required) but description only mentions connectionId. With 0% schema description coverage, this omission leaves workspaceId unexplained, reducing agent understanding.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool checks connector status after OAuth, with specific user queries. It distinctly differs from sibling tools like chieflab_connect_connector by focusing on status confirmation.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Explicitly specifies when to use ('just completed a connector OAuth flow') and gives example user questions. Could mention when not to use or alternatives, but provides strong context.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_connect_providerAInspect

Connect a third-party provider (Zernio, Resend, GA4, Search Console, HubSpot, Stripe, Linear, Notion, Slack) to this workspace. USE WHEN the user wants to wire up publishing, email sending, or analytics readback. For OAuth providers (ga4 / search_console / hubspot) returns an authorizeUrl the agent surfaces to the user. For API-key providers (zernio / resend) returns instructions for the set-key tool. Without this, publish/send/measure tools return 'configure first' errors.

ParametersJSON Schema
NameRequiredDescriptionDefault
providerYesProvider id: zernio | resend | ga4 | search_console | hubspot | stripe | linear | notion | slack
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Although no annotations exist, the description reveals that OAuth providers return an authorizeUrl and API-key providers return instructions for a set-key tool. It also mentions that without it, other tools return errors. However, it does not discuss whether it overwrites existing connections or other side effects.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Three sentences, no fluff. Each sentence serves a purpose: what it does, when to use, and expected behavior per provider type.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the simplicity (one parameter, no output schema), the description covers purpose, usage, behavioral differences, and prerequisites. It could mention idempotency or potential side effects, but overall it's adequate.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

With only one parameter and 100% schema description coverage, the description adds value by listing example providers and explaining behavioral differences (OAuth vs. API-key), which goes beyond the schema's basic enum.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool connects a third-party provider to the workspace, lists specific providers (Zernio, Resend, GA4, etc.), and distinguishes it from siblings that handle reading, publishing, or other actions.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines5/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

It explicitly says 'USE WHEN the user wants to wire up publishing, email sending, or analytics readback,' and differentiates between OAuth and API-key providers, including the expected return values. It also warns of consequences if not used before other tools.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_connect_publish_accountAInspect

USE WHEN the user wants to connect a social account (LinkedIn / X / Threads / Instagram / Facebook / Bluesky / TikTok) so chiefmo_publish_approved_post can post to it. Returns a Zernio OAuth URL — surface to user → they click → connect their account → it appears in chieflab_list_publish_accounts. Standard agent flow: call this with platform='linkedin' → get authUrl → surface to user → user connects → call chieflab_list_publish_accounts to confirm + grab accountId → pass accountId to chiefmo_publish_approved_post. NOTE: Zernio requires a profileId (the workspace container). If you don't pass one, the call fails with 'Profile ID required'. Get it from app.zernio.com/settings/profile, or call chieflab_list_publish_accounts first — the accounts response includes profileId on each item.

ParametersJSON Schema
NameRequiredDescriptionDefault
platformYeslinkedin | x | threads | instagram | facebook | bluesky | tiktok
profileIdNoZernio profile id. Required by Zernio's /connect endpoint. Get from app.zernio.com/settings/profile or from chieflab_list_publish_accounts response.
redirectUrlNoWhere Zernio should redirect after the user completes OAuth. Default: chieflab.io/runs (approval surface).
workspaceIdNoOptional workspace id (defaults to your API key's workspace).
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations provided, but description discloses tool returns OAuth URL, requires user interaction, and will fail with 'Profile ID required' if missing. Does not cover permissions or rate limits, but for a connect tool this is sufficient.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Description is front-loaded with purpose and includes a structured flow and note. Slightly long but every sentence adds value.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness5/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

No output schema, but description explains return value (OAuth URL), prerequisites, failure mode, and ties to sibling tool. Covers the entire user flow adequately.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

All 4 parameters have schema descriptions (100% coverage). Description adds value by explaining platform enum, profileId sourcing, and redirectUrl default, and the failure behavior for missing profileId.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool connects social accounts for posting, lists supported platforms, and distinguishes from sibling chieflab_list_publish_accounts by specifying the flow and output (OAuth URL).

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Explicitly says 'USE WHEN' and provides a standard flow, including how to obtain the required profileId. Names sibling tool for confirmation. Doesn't explicitly list when not to use, but context is clear.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_create_tenantAInspect

USE WHEN you (an AI product builder) are managing multiple end-users and want isolated brand context per user — 'create a tenant for my new customer', 'set up Acme Co as a tenant', 'multi-tenant brand isolation'. Each tenant gets its own brand context, voice samples, memory, scoped under your workspace.

ParametersJSON Schema
NameRequiredDescriptionDefault
nameYesDisplay name (e.g. 'Acme Co').
domainNoOptional URL/domain (used for auto-context extraction).
industryNoOptional industry (e.g. 'B2B SaaS').
tenantIdYesShort slug used in MCP calls (e.g. 'tenant_abc123' or 'acme-co').
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations provided, so description carries full burden. Mentions that each tenant gets its own brand context, voice samples, memory—important side effects. Lacks details on immediacy, reversibility, or authorization needs.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Concise single sentence plus usage prompt; no fluff. Front-loaded with 'USE WHEN' for quick identification.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Adequate for a creation tool with no output schema. Explains purpose and scope but could mention return value (e.g., created tenant object) or error conditions.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100%, so baseline 3. Description does not add meaningful extra details beyond schema field descriptions (e.g., no example values or usage hints).

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Clearly states the verb ('create') and resource ('tenant'), with context of isolating brand context per user. Differentiates from siblings like chieflab_list_tenants and chieflab_set_tenant_context.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Includes explicit 'USE WHEN' directive with example scenarios, guiding the agent to appropriate context. Lacks explicit when-not-to-use or alternative tools, but context is clear.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_customer_success_queueAInspect

Internal CS-ops queue: lists blocked launches, missing-approval items, recent connector failures, and due followups across the workspace. Use to triage what's stuck. Filterable by bucket.

ParametersJSON Schema
NameRequiredDescriptionDefault
limitNo
bucketNoblocked_launches | missing_approvals | connector_failures | due_followups | all (default)
workspaceIdNo
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations provided, so description must disclose behavioral traits. It only lists content types and filterability, but omits whether the operation is read-only, destructive, requires auth, or any side effects. Significant gap.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two sentences, no redundant words. Front-loaded with purpose and content. Every part earns its place.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

No output schema, no annotations, and incomplete parameter info. Description doesn't cover return format, pagination, or workspace scoping. Inadequate given tool's complexity.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters2/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema has 33% description coverage. Description adds context only for 'bucket' ('Filterable by bucket'), but leaves 'limit' and 'workspaceId' unexplained. Inadequate for 3-param tool with low schema coverage.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Description clearly states tool lists specific CS-ops items (blocked launches, missing approvals, etc.) and is differentiated from sibling tools like chieflab_launch_product or chieflab_list_tenants.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Explicitly says 'Use to triage what's stuck' and mentions filterability by bucket, providing clear usage context. Lacks explicit when-not-to-use or alternatives, but sufficient for its simplicity.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_get_connect_actionAInspect

P13 — for a single (typically blocked) channel, return the exact next tool to call to unblock it: chieflab_connect_provider for OAuth/key providers, chieflab_use_manual_fallback for manual-only channels. USE WHEN the user clicks an unblock button or the agent decides which channel to wire next.

ParametersJSON Schema
NameRequiredDescriptionDefault
channelYesChannel id from the catalog (publish_x / send_email / post_product_hunt / etc.) or a legacy alias (linkedin / email / product_hunt).
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description must disclose behavioral traits. It describes the tool as a lookup that returns a tool name, implying a read-only operation, but does not mention side effects, permissions, or error handling. This is adequate but leaves some gaps.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is extremely concise—two sentences that front-load the tool's purpose and usage context without extraneous information. Every sentence serves a clear purpose.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Despite having no output schema, the description adequately explains the tool's return value (the next tool to call) and its context (blocked channel). It covers the essential information for an agent to decide whether to use this tool.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100% for the single parameter 'channel'. The description mentions channel id examples but adds no new meaning beyond the schema's description. Therefore, baseline score of 3 applies.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool's function: returning the next tool to call to unblock a typically blocked channel. It explicitly names the two possible return values (chieflab_connect_provider and chieflab_use_manual_fallback), which distinguishes it from siblings.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides explicit usage context: 'USE WHEN the user clicks an unblock button or the agent decides which channel to wire next.' This gives clear guidance on when to invoke the tool, though it does not explicitly list cases when not to use it.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_get_users_after_buildAInspect

Outcome-named alias for chiefmo_launch_product. USE WHEN a coding agent just built, shipped, or deployed something and the user asks 'get users', 'find customers', 'make money from this', 'get paying users', 'get eyes on this', 'what now?', or 'how do I launch this?' This is the agent-dependency tool: after build, call this to turn the product into a repo-grounded money/distribution loop with launchPack, signed reviewUrl, approval-gated publish/send actions, 24h measurement, memory, next move, moneyLoop, and agentDependency. Gather repoContext first. Route exactly like chiefmo_launch_product; do not use chiefmo_diagnose_marketing for a new product.

ParametersJSON Schema
NameRequiredDescriptionDefault
goalNoDesired outcome, e.g. 'Get first 100 users', 'find paying customers', 'get design partners', 'make money from this'.
brandNoOptional inline brand context.
channelsNoOptional launch channels. Default: LinkedIn, X, Product Hunt, email, landing hero.
tenantIdNoOptional tenant/end-user id for stored brand context and memory.
outputModeNocontext | draft | full. Defaults to context/draft behavior from chiefmo_launch_product.
productUrlNoLive product URL. Pass either productUrl or productDescription.
repoContextNoStrongly recommended. Include whatChanged, recentCommits, changedFiles, routes, readme, targetCustomer, launchGoal.
idempotencyKeyNoOptional dedupe key to prevent duplicate launch preparation.
productDescriptionNoShort product description if no public URL exists yet.
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations provided, so description carries full burden. It details what the tool accomplishes: turning the product into a loop with launchPack, signed reviewUrl, approval-gated actions, 24h measurement, memory, next move, moneyLoop, and agentDependency. Also recommends prerequisite 'Gather repoContext first.' Lacks explicit mention of side effects like creation or mutation, but overall transparent.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Description is fairly long but front-loaded with purpose and usage. Each sentence adds value given the tool complexity. Could be slightly more concise but structure is logical and complete.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given 9 parameters, nested objects, and no output schema, the description covers prerequisites, usage triggers, outcomes, and guidance. It does not explain return values explicitly but mentions outputs like launchPack, reviewUrl, etc. Overall sufficient for the complexity.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters5/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100%, and description adds significant meaning: defaults for channels, recommended nature of repoContext, clarification of outputMode values ('context | draft | full'), and relationship between productUrl and productDescription. Provides context beyond schema.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Explicitly states it is an outcome-named alias for chiefmo_launch_product, lists specific user queries that trigger it, and distinguishes from sibling tools like chiefmo_diagnose_marketing. The verb 'turn the product into' plus resource 'repo-grounded money/distribution loop' provides a specific purpose.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines5/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Provides explicit when-to-use scenarios: 'USE WHEN a coding agent just built, shipped, or deployed something and the user asks...' and includes a list of example user queries. Also gives routing guidance: 'Route exactly like chiefmo_launch_product; do not use chiefmo_diagnose_marketing for a new product.'

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_launch_productAInspect

Operator: chieflab-launch (primary). Spec v0.1 reference implementation of the orchestrator. End-to-end launch loop: repo-grounded positioning + per-channel drafts (linkedin, x, hn, reddit, email, landing_hero, product_hunt) + image briefs + approval-gated publishActions + signed reviewUrl + 24h measurement queue. Internally composes chieflab-post + chieflab-email under one runId. Spec-renamed alias of chiefmo_launch_product (backwards-compat retained).

ParametersJSON Schema
NameRequiredDescriptionDefault
goalNo
brandNo
channelsNo
tenantIdNo
productUrlNo
webhookUrlNo
repoContextNo
workspaceIdNo
imagesNeededNo
idempotencyKeyNo
Behavior5/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description bears full responsibility and does well by detailing internal composition of chieflab-post and chieflab-email, approval-gated publishActions, signed reviewUrl, and 24h measurement queue.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness3/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single dense paragraph packing much information, but it lacks structure like bullet points or clear separation. It is moderately concise but could be better organized for readability.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Despite describing the high-level orchestration process, the tool has 10 parameters, nested objects, and no output schema. The description does not cover parameter usage or return values, leaving the agent underinformed for a complex tool.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters1/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The description provides no explanation of any of the 10 input parameters. Given 0% schema coverage, the agent receives no guidance on parameter meaning or usage, leaving a significant gap.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states it is an orchestrator for an end-to-end launch loop, listing specific channels and steps. It distinguishes itself from the sibling tool chiefmo_launch_product by noting it is a spec-renamed alias, providing unique identity.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description identifies it as the 'primary' operator and a reference implementation, giving implicit context for usage. However, it lacks explicit when-to-use or when-not-to-use guidelines compared to other sibling tools.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_list_available_channelsAInspect

P13 — list every channel ChiefLab can ship to (LinkedIn / X / Threads / Bluesky / Instagram / Facebook / TikTok / Product Hunt / Hacker News / Reddit / Discord / Slack / Indie Hackers / DEV.to / YC Bookface / email / blog / GA4 / Search Console / Stripe). Returns the catalog grouped by category (social / community / email / blog / analytics) with each channel's capabilities (publish / schedule / read_metrics / connect / fallback) and providers. USE WHEN the user asks 'what channels can ChiefLab post to?', 'which platforms are supported?', or before calling chiefmo_launch_product to pick the right channels.

ParametersJSON Schema
NameRequiredDescriptionDefault
categoryNoOptional. Filter to one category: social | community | email | blog | analytics.
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations provided, so description carries burden. It describes a read operation without side effects, but does not explicitly state non-destructive nature or any other behavioral traits.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Description is well-structured: first line states purpose, then output format, then usage examples. Every sentence adds value without redundancy.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness5/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given simple tool with one optional parameter and no output schema, description fully explains what the tool returns (grouped catalog with capabilities and providers) and when to use it.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100% and description adds context by listing categories and explaining that category parameter filters the result, enhancing understanding beyond schema.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Description clearly states it lists available channels, enumerates many examples, and explains output grouping and capabilities. Distinguishes from sibling tools by focusing on channel catalog rather than other operations.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines5/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Explicitly provides trigger queries and recommends use before calling chiefmo_launch_product, giving clear when-to-use guidance.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_list_channel_readinessAInspect

P13 — for every channel in the catalog, return its readiness state for THIS workspace: { ready, provider, blocker, fix, manualFallback, fallbackKey }. Drives the dashboard connector-readiness panel and the launch-time blockedActions list. USE WHEN the user asks 'which channels can I publish to right now?' or before chiefmo_launch_product to pre-warn about blocked channels.

ParametersJSON Schema
NameRequiredDescriptionDefault
workspaceIdNoOptional workspace id.
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations exist, so description carries full burden. It explains output format and usage context (drives dashboard panels), but lacks details on permissions, rate limits, or side effects. Since it's a read operation, the description is adequate but not detailed.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is extremely concise—two sentences. First sentence defines output keys and scope, second gives usage guidance. No wasted words or repetition.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given no output schema, the description explains return format sufficiently. It also provides usage context. However, it does not mention potential edge cases like empty catalog or error conditions, which would make it more complete.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100% with a single optional workspaceId parameter described as 'Optional workspace id.' The description does not add further semantics about default behavior or constraints, so baseline 3 is appropriate.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool returns readiness state for every channel in the catalog for the current workspace, using specific terms like ready, provider, blocker, etc. It distinguishes purpose from siblings like chieflab_list_available_channels by focusing on readiness.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description explicitly tells when to use: when user asks about channels they can publish to right now or before chiefmo_launch_product. It provides concrete query examples, though it doesn't list when not to use or explicitly name alternatives.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_list_connectorsAInspect

USE WHEN you want to know which third-party connectors are connected for this workspace — 'what data sources do I have wired?', 'is X connector live?'. Returns all connectors across all operators with OAuth-live status (live data vs. mock snapshots).

ParametersJSON Schema
NameRequiredDescriptionDefault

No parameters

Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description discloses that it returns all connectors across operators with OAuth-live status, distinguishing live data from mock snapshots. It does not mention read-only nature or permission requirements, but the behavioral context is good for a list operation.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two sentences with clear front-loading: first sentence states purpose, second specifies returns. No wasted words.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness5/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a simple list tool with no parameters and no output schema, the description fully explains what it returns (all connectors with OAuth-live status) and scope (per workspace, across all operators). Complete for its complexity.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The input schema is empty, so schema coverage is 100% trivially. The description adds no parameter info (none needed). Baseline 4 is appropriate for a parameterless tool.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool lists third-party connectors connected to the workspace, with example queries. It distinguishes from siblings like chieflab_connector_status (specific status) and chieflab_list_available_channels (available channels) by explicitly focusing on connectors with OAuth-live status.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description explicitly says 'USE WHEN you want to know which third-party connectors are connected' and provides sample queries. However, it does not mention when not to use or list alternatives among siblings, so it lacks exclusions.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_list_email_sendersAInspect

USE WHEN you need to know which from addresses chiefmo_send_approved_email will accept — 'list my verified email domains', 'which sender can I use?'. Returns verified Resend sender domains for the workspace. Requires RESEND_API_KEY (env-default or set via chieflab_set_resend_key).

ParametersJSON Schema
NameRequiredDescriptionDefault
workspaceIdNoOptional workspace id.
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Description is a read operation but does not explicitly state non-destructive behavior or side effects. With no annotations, could be improved by clarifying it only lists data and does not modify anything.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two sentences, front-loaded with usage context, no extraneous words. Every sentence serves a purpose.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given simple tool with one optional param and no output schema, description covers prerequisite (API key), return value, and usage context. Lacks only a mention of response format, which is acceptable.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100% for the single optional parameter. Description adds context that it returns data for the workspace, but does not significantly enhance meaning beyond schema.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Description specifies it lists verified Resend sender domains (from addresses) for the workspace, with example user queries. Distinguishes from sibling tools like chieflab_set_resend_key.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

States explicit usage trigger: 'USE WHEN you need to know which from addresses chiefmo_send_approved_email will accept'. Also mentions prerequisite of RESEND_API_KEY, providing clear context.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_list_launch_templatesBInspect

List launch templates saved for this workspace (most recently created first). Pass label of one to chiefmo_launch_product as templateId for a one-call repeat launch.

ParametersJSON Schema
NameRequiredDescriptionDefault
limitNo
workspaceIdNo
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description must fully disclose behavior. It mentions sorting but omits that this is a read-only operation, does not mention error handling, pagination, or authentication requirements.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two concise, front-loaded sentences communicate the core purpose and a key usage hint without any redundant information.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the lack of output schema and annotations, the description does not explain the return format, parameter details, or error behavior, making it insufficient for reliable agent use.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters1/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 0%, yet the description provides no explanation of either parameter (limit, workspaceId), leaving the agent to infer their meaning solely from the schema.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool lists launch templates for the workspace, sorted by most recently created, and distinguishes it from sibling tools like chieflab_save_launch_template by indicating it is for reading lists.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description explains when to use this tool (to list templates) and provides an explicit usage scenario: passing a label to chiefmo_launch_product. However, it does not explicitly state when not to use it or alternative tools for other purposes.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_list_playbooksAInspect

List the ChiefLab launch playbooks (devtool / saas / ai_agent / marketplace / internal_tool / consumer) — what's the default channel mix, default angle, ICP, proof expectations, risks-to-avoid for each. Use to show the user options before calling chiefmo_launch_product, or to pick productType explicitly when auto-detection is wrong.

ParametersJSON Schema
NameRequiredDescriptionDefault

No parameters

Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations provided. Description explains it lists playbooks with details, implying read-only behavior, but does not explicitly state non-destructive nature or other traits. Adequate but not rich.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Single, well-structured sentence with essential information front-loaded. No wasted words.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness5/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given no output schema, description fully explains what the tool returns: list of playbook types and their attributes. Complete for its intended use.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Input schema has zero parameters, so description adds no parameter info. Baseline is 4 for no parameters.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states that the tool lists ChiefLab launch playbooks with specific types (devtool, saas, etc.) and details (channel mix, angle, ICP). It distinguishes itself from the sibling tool 'chiefmo_launch_product' by indicating its use before that tool.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines5/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Explicitly states when to use: 'before calling chiefmo_launch_product' and 'to pick productType explicitly when auto-detection is wrong'. Provides clear context and alternatives.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_list_publish_accountsAInspect

USE WHEN you need accountIds before publishing — 'which social accounts are connected?', 'list my LinkedIn/X/Instagram accounts'. Returns connected Zernio accounts (LinkedIn, X, Threads, Instagram, Facebook, etc.) with the accountIds you'll pass to chiefmo_publish_approved_post. Requires ZERNIO_API_KEY in env or via chieflab_set_zernio_key.

ParametersJSON Schema
NameRequiredDescriptionDefault
workspaceIdNoOptional. Defaults to the workspace derived from your API key.
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description carries the full burden. It discloses that the tool returns connected accounts and accountIds, implying a read operation. It also notes the API key prerequisite. Missing details on pagination, error handling, or idempotency, but adequate for a simple list tool.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two sentences, front-loaded with 'USE WHEN', no redundant information. Every sentence is essential and well-structured.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the simple tool (one optional param, no output schema, no nested objects), the description provides necessary context: what it returns, why it's used, and a prerequisite. Lacks mention of edge cases like no connected accounts, but minor.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The input schema has 100% description coverage for the one optional parameter workspaceId. The description does not add additional meaning about this parameter beyond the schema; it focuses on the purpose of the output. Baseline of 3 is appropriate.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the verb 'list' and resource 'publish accounts', provides example user queries, and distinguishes from any sibling tool by specifying it returns accountIds for publishing. It is specific and unambiguous.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Explicitly says 'USE WHEN you need accountIds before publishing', giving clear context. It also mentions the API key requirement. However, it does not explicitly state when not to use this tool or mention alternatives among siblings, which is a minor gap.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_list_tenantsAInspect

USE WHEN you want to see which end-users (tenants) are configured in this workspace — 'list my tenants', 'show me all customers I've onboarded'. Returns active tenants by default; pass status='paused'|'archived' for others.

ParametersJSON Schema
NameRequiredDescriptionDefault
statusNoactive | paused | archived (default: active)
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description discloses that the tool is a read operation returning active tenants by default and how to get other statuses. It does not discuss rate limits or permissions, but for a simple list tool, this is sufficient.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two sentences with zero wasted words. First sentence gives purpose and usage patterns; second specifies defaults and filter options. Highly efficient.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

The description covers the tool's purpose, default behavior, and parameter use. For a simple list tool with no output schema, it is mostly complete, though it could mention the response format (e.g., names and IDs).

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The only parameter 'status' is fully described in the schema (100% coverage). The description adds natural language examples and restates the default, but does not provide significant new meaning beyond the schema.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Description clearly states the tool lists tenants in the workspace, backed by natural language examples ('list my tenants', 'show me all customers I've onboarded'). It distinguishes from sibling tools like chieflab_create_tenant and other list tools.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Provides explicit usage context ('USE WHEN you want to see...') and explains default behavior. However, it does not mention when not to use it or suggest alternative sibling tools for related tasks.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_measureBInspect

Operator: chieflab-measure (primary). 24h post-launch readback: pulls Zernio engagement, GA4 traffic, Search Console queries; returns structured metrics + next-iteration brief. Spec-renamed alias of chiefmo_post_launch_review — both names route to the same handler.

ParametersJSON Schema
NameRequiredDescriptionDefault
runIdYes
workspaceIdNo
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description must cover behavioral traits. It explains that the tool reads data and returns metrics/brief, implying a read operation. However, it does not mention permissions, rate limits, or any potential side effects, leaving some gaps.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is short (two sentences) and front-loaded with key information. It is mostly concise but includes some jargon ('Zernio engagement') that may not be universally understood.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the complexity (no output schema, 2 parameters), the description is incomplete. It fails to explain what the returned 'structured metrics' include or the meaning of the input parameters. The agent needs more context to use the tool correctly.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters1/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The description provides no information about the parameters (runId, workspaceId) beyond what is in the schema. With 0% schema coverage, this is a significant omission. The agent cannot infer parameter semantics from the description.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description explicitly states the tool's function: '24h post-launch readback' that pulls data from specific sources (Zernio, GA4, Search Console) and returns structured metrics plus a brief. It also clarifies it is a renamed alias of chiefmo_post_launch_review, distinguishing it from siblings.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description indicates the tool is for '24h post-launch readback,' providing clear timing context. However, it lacks explicit guidance on when not to use it or alternatives beyond the alias mention.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_postAInspect

Operator: chieflab-post (primary). Single-channel publish loop. Accepts channel + product context, returns one draft + one publishAction + reviewUrl. Use when the user says "post this to linkedin" / "draft a hacker news submission" / "send a tweet about this" — anywhere the full launch loop is overkill. Approval-gated, spec v0.1 conformant. Falls through to the launch handler with channels=[channel], imagesNeeded=0.

ParametersJSON Schema
NameRequiredDescriptionDefault
goalNo
brandNo
channelYeslinkedin | x | hacker_news | reddit | product_hunt | threads | instagram
tenantIdNo
productUrlNo
repoContextNo
workspaceIdNo
idempotencyKeyNo
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations exist. Description discloses 'approval-gated' and 'spec v0.1 conformant' but lacks details on side effects, authentication needs, rate limits, or data mutation beyond publishing.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness3/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Description includes extraneous phrases like 'Operator: chieflab-post (primary)' and 'spec v0.1 conformant' that add little value. Could be more concise while retaining key info.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given 8 parameters, no output schema, and no annotations, description covers purpose and usage context but omits parameter details and return value format. Leaves gaps for an agent to infer.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters2/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 13% (only 'channel' described). Description references 'channel + product context' but does not explain other 6 parameters (goal, brand, tenantId, productUrl, repoContext, workspaceId, idempotencyKey). Incomplete.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Description clearly identifies tool as a single-channel publish loop, accepts channel and product context, returns draft/publishAction/reviewUrl. Distinguishes from sibling 'chieflab_launch_product' by calling it 'full launch loop overkill'.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines5/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Explicit usage examples: 'post this to linkedin', 'draft a hacker news submission', 'send a tweet about this'. States when not to use ('full launch loop is overkill') and hints at fallback to launch handler.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_record_manual_publishAInspect

USE WHEN the user has manually posted to a channel returned by chieflab_use_manual_fallback (Product Hunt / HN / Reddit / Discord / etc.) and wants to feed the live URL back to ChiefLab so the closed loop continues. Records the URL on the original publishAction (status flips from 'approved' to 'executed_manually'), persists a proof_asset to the P9 company brain, and queues 24-hour metrics readback via chiefmo_post_launch_review. Without this tool, manually-posted channels are lost to ChiefLab's measurement loop.

ParametersJSON Schema
NameRequiredDescriptionDefault
notesNoOptional. Any context on how the post performed in the first hour (early upvotes, top comment, reactions). Persists into the P9 brain.
actionIdYesThe publishAction.id that the user manually published (from chieflab_use_manual_fallback). Required.
publishedAtNoOptional ISO 8601 timestamp. Defaults to now.
workspaceIdNoOptional workspace id.
publishedUrlYesThe live URL of the manual post (e.g. producthunt.com/posts/<slug>, news.ycombinator.com/item?id=<id>, reddit.com/r/SaaS/comments/<id>).
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full burden of disclosing behavioral traits. It describes the three main effects (status update, proof persistence, metrics queuing) but omits details on authorization, idempotency, or rate limits. The description is adequate but not comprehensive.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is well-structured with a leading 'USE WHEN' statement. It is concise but covers necessary details in a few sentences. Could be slightly shorter, but no redundant sentences.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given no output schema, the description does not explain return values. It does adequately describe the side effects (status change, proof asset, metrics queue). With 5 parameters and high schema coverage, the description is mostly complete but lacks output documentation.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100%: all 5 parameters have descriptions in the input schema. The description adds context by explaining the purpose of 'actionId' and 'publishedUrl' and notes that 'notes' persist into the P9 brain, but this information largely overlaps with the schema descriptions. Baseline 3 applies since schema already covers parameters well.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly identifies the specific action: recording a manual publish URL back to ChiefLab. It details the exact operations (status flip, proof persistence, metrics queuing) and distinguishes from sibling tools like chieflab_use_manual_fallback and other record tools.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description begins with 'USE WHEN' and explicitly states the prerequisite (manual post to a channel from chieflab_use_manual_fallback). It explains why the tool is necessary ('Without this tool, manually-posted channels are lost'), providing clear context. However, it does not mention when not to use the tool or alternative scenarios.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_record_rendered_copyAInspect

USE WHEN your agent's LLM has rendered the final copy from a launch-pack brief and you want the human reviewer to see the rendered text on the reviewUrl (not the brief). Closes the orchestration loop: ChiefLab returns briefs (cheap), your LLM renders (high-quality), this tool writes the rendered copy back so reviewers approve real text, not prompts. Pass the assetId from launchPack.channels[channel].assetId (or any other ChiefLab asset). Idempotent — second call replaces the prior rendered body.

ParametersJSON Schema
NameRequiredDescriptionDefault
bodyYesThe final rendered copy your LLM produced. Markdown is fine. Replaces the brief on the reviewUrl.
assetIdYesThe asset to attach rendered copy to. Get from launchPack.channels[channel].assetId or run.outputs[i].assetId.
renderedByNoOptional. Which model produced this (e.g. 'sonnet-4.6', 'gpt-4o', 'gemini-2.5-pro'). Stored in metadata for audit.
workspaceIdNoOptional workspace id.
Behavior5/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Despite no annotations, the description discloses key behaviors: it closes the orchestration loop, is idempotent (replaces prior rendered body on second call), and writes to the reviewUrl so reviewers see real text. No contradictions.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Description is somewhat long but each sentence adds value: use case, orchestration context, parameter guidance, idempotency. Could be slightly trimmed but remains focused. No wasted sentences.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Tool is simple (4 params, 2 required, no output schema). Description explains idempotency and where to find assetId. For a write operation, this covers the essential context. Could mention potential error states, but not critical.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100%, but description adds value by explaining the purpose of each parameter beyond the schema: body is 'final rendered copy', assetId source is clarified, renderedBy for audit, workspaceId optional. This adds meaningful context.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Description clearly states the tool writes rendered copy from a launch-pack brief for human review. It specifies the verb 'writes' and the resource 'reviewUrl', and distinguishes from siblings like chieflab_brain_read by explaining the orchestration loop.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description explicitly provides a use case ('USE WHEN your agent's LLM has rendered the final copy...') and explains the orchestration flow. However, it does not explicitly mention when not to use this tool or list alternative tools for different scenarios.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_record_voice_sampleAInspect

USE WHEN the user approves or rejects a draft and you want the next run to learn — 'remember this is the right tone', 'never write like this again', 'add this email as a reference'. Stores approved/rejected examples per tenant; future operator prompts ground against approved tone and avoid rejected patterns. The 'rejection becomes memory' loop.

ParametersJSON Schema
NameRequiredDescriptionDefault
kindYesapproved | rejected | reference
channelNoemail | social | landing | ad | blog (optional)
contentYesThe actual text
feedbackNoOptional rejection feedback like 'too corporate, make it more founder-led'
tenantIdYes
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Discloses that data is stored per tenant and that future prompts will ground against approved tone and avoid rejected patterns. The 'rejection becomes memory' loop is described. No annotations, so description carries burden well; could still mention if data is appended or overwritten.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Description is front-loaded with usage condition, then explains mechanism. It is informative but slightly verbose; every sentence adds value, but could be tightened.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Covers main functionality and outcome (learning loop). Lacks details on error handling or multiple calls behavior, but for a data recording tool, it is sufficiently complete for an agent to decide when to use.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Description adds meaning to parameters by explaining the kind values (approved/rejected/reference) and the purpose of feedback field. Schema coverage is 80% but descriptions are brief; this context helps the agent understand usage beyond schema.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Description clearly identifies the tool as storing approved/rejected voice samples per tenant to influence future generations. It uses specific verbs and distinguishes the purpose from sibling tools like chiefmo_record_brand_voice by focusing on feedback loop.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Explicit 'USE WHEN' triggers (user approves/rejects draft) provide clear guidance. However, no explicit mention of when not to use or alternatives among siblings, though context implies specific scenario.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_remember_preferred_channelAInspect

P13 — record a workspace's per-job preference (e.g. 'this workspace prefers LinkedIn over X for B2B launches', 'always use the founder LinkedIn account, not company page'). Future runs of chiefmo_launch_product read these preferences before picking the channel mix. Persists into the P9 brain.

ParametersJSON Schema
NameRequiredDescriptionDefault
channelYesChannel id (publish_linkedin, send_email, etc.).
preferenceYesFree-form preference object (e.g. { accountId, persona, weight }).
workspaceIdNo
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Without annotations, the description carries the full transparency burden. It discloses persistence ('persists into the P9 brain') and downstream reading, but does not clarify overwrite behavior (append vs replace), required authorization, or concurrency effects, leaving gaps for an agent.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is two concise sentences, front-loaded with the key action and examples. Every sentence adds value—usage context and persistence. No redundancy or wasted words.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool has 3 parameters (one nested), no output schema, and no annotations, the description explains purpose and downstream impact but omits return value/confirmation, error handling, and behavior on repeated calls. Adequate but with notable gaps.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is high (67% with descriptions), so the baseline is 3. The tool description provides examples of preference content but does not add structural detail beyond the schema's description of the 'preference' object or clarify the optional workspaceId parameter meaningfully.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description explicitly states the action ('record a workspace's per-job preference') with concrete examples ('always use the founder LinkedIn account'), clearly identifying the verb and resource. It distinguishes from sibling tools like chiefmo_launch_product and brain_read by focusing on writing preferences.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description explains that the recorded preference influences chiefmo_launch_product, providing clear context for when to use the tool. However, it does not specify when not to use it or mention alternative tools (e.g., chieflab_set_tenant_context) that might serve different purposes, slightly reducing guidance.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_retry_blocked_actionAInspect

P13 — re-check a blocked publishAction's connector readiness. Returns whether the channel is now ready (so the agent can re-fire chiefmo_publish_approved_post / chiefmo_send_approved_email) or still blocked with the up-to-date connect path. USE WHEN the user has just connected a missing provider and the agent needs to know if the action can fire.

ParametersJSON Schema
NameRequiredDescriptionDefault
actionIdYesThe blocked publishAction.id from chiefmo_launch_product.
workspaceIdNo
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations provided, so description carries full burden. It discloses that the tool checks readiness, returns whether ready or still blocked with connect path. It implies a non-destructive check. It does not mention permissions, side effects, or rate limits, but for a simple retry check it is adequate.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is concise at 2-3 sentences, front-loaded with the main purpose, and structured logically: action, return, usage condition. Every sentence adds value with no redundancy.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

The description covers return values (ready or still blocked with path) and suggests next steps. With no output schema, this is helpful. However, it lacks mention of prerequisites (e.g., action must exist) or error cases. Given the tool's simplicity, it is fairly complete.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters2/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The description adds no parameter information beyond the schema. The schema describes `actionId` but not `workspaceId`. With 50% schema description coverage, the description should compensate for the missing `workspaceId` description but fails to do so. No additional meaning or format details are provided.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool's purpose: 're-check a blocked publishAction's connector readiness.' It specifies the resource (blocked publishAction) and verb (re-check). It distinguishes from siblings by being specific to retrying a blocked action after a provider connection, not just general connector status.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description explicitly says 'USE WHEN the user has just connected a missing provider and the agent needs to know if the action can fire.' This provides clear context for usage. It also suggests follow-up actions (re-fire specific tools). However, it does not explicitly state when not to use or list alternatives.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_save_launch_templateAInspect

Save a launch's shape (channels + angle + brand + default repo context) as a reusable template. Use AFTER a launch performs well so the next one in the same shape skips re-asking. The current run's productUrl and assets are NOT saved — just the orchestration shape.

ParametersJSON Schema
NameRequiredDescriptionDefault
icpNo
brandNo
labelYesShort label, e.g. 'Devtool launch — bottom-up'.
channelsNo
descriptionNoOptional one-line description of when to use this template.
launchAngleNo
productTypeNoOne of: devtool | saas | ai_agent | marketplace | internal_tool | consumer.
workspaceIdNo
sourceLaunchIdNoOptional launchId to copy channels/angle/brand from. Without it, pass the fields explicitly.
defaultRepoContextNo
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description carries the full burden. It discloses that productUrl and assets are not saved (exclusions) and implies mutation. However, it does not detail side effects like overwriting existing templates, permission requirements, or return values, leaving moderate gaps.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Three sentences with no fluff. The first sentence states purpose and content, the second gives usage timing, and the third clarifies exclusions. Every sentence adds value, making it efficient and front-loaded.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the complexity (10 parameters, nested objects, no output schema, no annotations), the description covers core purpose, exclusions, and usage timing. However, it omits details about required vs. optional parameters (only label is required) and does not elaborate on nested objects like icp or brand, leaving some gaps for a complete understanding.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 40%, with some parameters having descriptions (label, description, sourceLaunchId, productType). The description clarifies the purpose of channels, angle, brand, and defaultRepoContext by labeling them as 'shape.' For the many undocumented parameters (icp, brand, workspaceId, etc.), the description provides indirect meaning but lacks explicit field-level detail, setting a baseline of 3.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool saves a 'launch's shape' (channels, angle, brand, default repo context) as a reusable template. It specifies what is excluded (productUrl, assets), providing a precise verb+resource. This distinguishes it from related siblings like chieflab_list_launch_templates.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description explicitly advises to use this tool 'AFTER a launch performs well so the next one in the same shape skips re-asking,' giving clear context. It does not mention when not to use or provide alternatives, but the guidance is sufficient for typical scenarios.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_score_launch_packAInspect

Re-score a launch pack on six quality dimensions (specificity / claim_risk / channel_fit / icp_clarity / proof / readiness). Use AFTER chieflab_record_rendered_copy populates rendered bodies — the original chiefmo_launch_product call returned a score for the briefs, but rendered copy needs its own quality check before approval.

ParametersJSON Schema
NameRequiredDescriptionDefault
runIdYesThe launchId from chiefmo_launch_product. ChiefLab pulls the latest assets + actions for this run and re-scores.
workspaceIdNoOptional workspace id.
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full burden. It discloses the quality dimensions and the prerequisite, but does not clarify whether the tool mutates data (writes scores) or is read-only, nor does it mention permissions or response format.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, well-structured sentence that front-loads the key action and dimensions, then provides necessary context without redundancy.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

The description covers purpose, usage context, and prerequisites, but lacks information about return values (no output schema) and potential side effects. For a tool that re-scores, the output is important context.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100%; both parameters are documented in the schema. The description adds useful context for runId (source and effect) but workspaceId is only restated. Baseline is 3, and the added value is marginal.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool's purpose: re-scoring a launch pack on six named quality dimensions. It distinguishes itself from sibling tools by specifying the prerequisite (chieflab_record_rendered_copy) and contrasting with the earlier score from chiefmo_launch_product.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description explicitly states when to use this tool: 'Use AFTER chieflab_record_rendered_copy populates rendered bodies' and explains why the earlier score is insufficient. It provides clear context and ordering but does not explicitly state when not to use it.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_sendAInspect

Operator: chieflab-email (primary executor). Approval-gated email send via Resend. Pass actionId from a prior launch's publishActions (after the user approves the email). Spec-renamed alias of chiefmo_send_approved_email — both names route to the same handler.

ParametersJSON Schema
NameRequiredDescriptionDefault
runIdNo
actionIdYes
idempotencyKeyNo
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description carries the full burden. It discloses that the tool is approval-gated and an alias, but does not detail error handling, idempotency, or return behavior. Adequate but not comprehensive.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is concise at two sentences, front-loading the key purpose and usage. Every sentence adds value with no fluff.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given no output schema, no annotations, and 0% schema description coverage, the description does not cover the full context. Missing details on runId, idempotencyKey, return values, and failure modes make it incomplete for a tool with three parameters.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters2/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 0%, so the description must compensate. It only explains actionId's origin (publishActions) but does not mention runId or idempotencyKey. This leaves significant gaps for the agent.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool's purpose: an approval-gated email send via Resend. It specifies the verb 'send', the resource 'email', and the approval context. It also distinguishes itself from siblings by noting it is an alias of chiefmo_send_approved_email.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides clear guidance on when to use this tool: after user approval, passing actionId from a prior launch's publishActions. It implicitly tells the sequence but does not explicitly state when not to use it or list alternatives.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_set_resend_keyAInspect

USE WHEN the user wants to bring their own Resend account (per-workspace email sending) instead of the platform default. Stores the Resend API key encrypted, then chiefmo_send_approved_email + chieflab_list_email_senders use this key for the workspace. Get a Resend API key from resend.com/api-keys.

ParametersJSON Schema
NameRequiredDescriptionDefault
apiKeyYesYour Resend API key (re_... format). Stored encrypted at rest.
workspaceIdNoOptional workspace id.
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description reveals that the API key is 'stored encrypted at rest', which is important for security. However, it doesn't disclose whether setting a new key overwrites the old one, rate limits, or other side effects.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is concise with three sentences, front-loading the usage condition ('USE WHEN'). Every sentence adds value without redundancy.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given no annotations, no output schema, and simple parameters, the description covers purpose, usage, and key behavioral detail (encrypted storage). It could mention error handling or overwrite behavior, but overall it's sufficient.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100%, so the schema already documents both parameters. The description adds limited extra meaning: for apiKey it mentions 'encrypted at rest', and for workspaceId it says 'optional'. This adds some value but is not extensive.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the action ('stores the Resend API key encrypted') and the usage scenario ('USE WHEN the user wants to bring their own Resend account'). It distinguishes itself from sibling tools by specifying the purpose related to per-workspace email sending.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Explicitly starts with 'USE WHEN', providing clear context for when to use this tool. It mentions dependent tools (chiefmo_send_approved_email, chieflab_list_email_senders), but doesn't explicitly state when not to use or provide alternative tool names.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_set_tenant_contextAInspect

USE WHEN onboarding a new tenant or refreshing their brand — 'set up brand context for tenant X', 'update Acme Co's voice and pillars'. Upsert: stores brand name, audience, voice, strategic pillars, competitors, content angles, positioning risks. Every operator output for this tenant grounds against this context. Pair with chieflab_create_tenant.

ParametersJSON Schema
NameRequiredDescriptionDefault
brandNoBrand name
voiceNoTone of voice + brand personality
pillarsNo3-5 strategic pillars
audienceNo
categoryNo
tenantIdYes
competitorsNo
contentAnglesNo
positioningRisksNo
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Description discloses upsert behavior and states that operator outputs will ground against this context. However, with no annotations, it should detail merge behavior on update, permissions, or side effects. The description is adequate but leaves gaps.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Description is front-loaded with usage instructions and examples. It is clear and efficient, though the examples could be slightly condensed. Overall, well-structured for an AI agent.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a tool with 9 parameters, no output schema, and no annotations, the description provides essential purpose and impact but lacks return value description, error conditions, or parameter validation rules. It meets minimum viability.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is only 33%, but the description lists most relevant fields (brand, audience, voice, pillars, competitors, content angles, positioning risks) and explains their role as brand context. It adds meaning beyond the schema but does not detail each parameter's specific use or constraints.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Description starts with a specific use case ('onboarding a new tenant or refreshing their brand') and gives example queries. It clearly identifies the action (upsert brand context) and resource (tenant context). It also distinguishes from sibling tool chieflab_create_tenant by suggesting pairing.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Explicitly states when to use ('USE WHEN onboarding a new tenant or refreshing their brand') and provides example natural language prompts. It recommends pairing with chieflab_create_tenant, but lacks explicit when-not-to-use or alternative tools for similar tasks.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_set_workspace_memberBInspect

Add or update a workspace member's role (owner | approver | editor | viewer). Returns the updated member record.

ParametersJSON Schema
NameRequiredDescriptionDefault
roleYesowner | approver | editor | viewer
emailNo
userIdYesThe user identifier (typically Supabase auth.uid or email).
workspaceIdNo
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description carries full burden. It discloses that the operation adds or updates and returns the updated member record, but lacks details on side effects (e.g., idempotency, error cases, permission requirements).

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is two sentences, front-loaded with the action and result. No unnecessary words or redundancy.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool has 4 parameters, no output schema, and no annotations, the description is somewhat complete for a simple mutation but lacks error handling, role validation details, and prerequisites. With siblings like chieflab_workspace_members, the mutation intent is clear.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters2/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 50%, with descriptions for role and userId only. The description mentions the allowed roles, which is already in the schema, but does not explain email or workspaceId. It adds minimal value beyond the schema.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the action (Add or update) and the resource (workspace member's role), listing the allowed roles. This distinguishes it from sibling tools like chieflab_workspace_members, which likely lists members.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No explicit guidance on when to use this tool vs alternatives, such as when to add vs update, or prerequisites like workspace existence. The description is purely functional.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_set_zernio_keyAInspect

USE WHEN the user wants to bring their own Zernio account (per-workspace social publishing) instead of using the platform default. Stores the Zernio API key encrypted, then chieflab_list_publish_accounts / chieflab_connect_publish_account / chiefmo_publish_approved_post / chiefmo_post_launch_review all use this key for the workspace. Get a Zernio API key from zernio.com.

ParametersJSON Schema
NameRequiredDescriptionDefault
apiKeyYesYour Zernio API key (sk_live_... or similar). Stored encrypted at rest.
workspaceIdNoOptional workspace id.
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

The description discloses that the key is 'stored encrypted at rest', which is a useful behavioral trait. However, with no annotations provided, the description could have covered more details such as error handling, idempotency, or authorization requirements. The encryption note adds value but leaves room for improvement.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Three sentences, each with a clear purpose: usage condition, dependent tools, and API key source. Front-loaded with imperative instruction. No redundant or unnecessary text.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness5/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a simple two-parameter tool with no output schema, the description covers when to use, what it does, how it relates to other tools, and how to get the key. It is complete enough for an agent to correctly select and invoke the tool.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100% and the description adds little beyond the schema: it mentions the API key format (sk_live_...) but that is already in the schema's description. The optional workspaceId is not elaborated. Baseline is 3 due to high schema coverage, and the description provides marginal extra insight.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool's purpose: 'USE WHEN the user wants to bring their own Zernio account' and specifies it stores the API key encrypted. It identifies the resource (Zernio API key) and the action (store), distinguishing it from other config tools on the server.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description explicitly states when to use the tool with 'USE WHEN' and lists dependent tools that require this key, providing clear context. It also tells users where to obtain the API key. However, it does not mention when not to use it or alternative approaches, leaving some gap.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_signup_workspaceAInspect

USE WHEN the user has no ChiefLab API key yet and you've gotten a 401 / 'authentication required' error from any other tool. Agent-first signup: creates a new workspace and returns the API key inline as apiKey. PREFERRED flow: use your filesystem tools to write the apiKey into the user's MCP config (see mcpConfigSnippet for the patch shape), then ask the user to restart their runtime once. After restart, re-call the original tool. FALLBACK: if you can't write to the config file, surface the included deliveryUrl to the user — they click, see the key, paste it manually. URL expires in 1 hour, single-use, IP-rate-limited (5/IP/hr). No login form.

ParametersJSON Schema
NameRequiredDescriptionDefault
agentNameNoOptional: which AI agent is requesting (Cursor, Claude Desktop, Codex, custom). Logged for support.
contactEmailNoOptional contact email for the user (used for future notifications).
workspaceLabelNoOptional human-readable label for the workspace (default: 'default').
Behavior5/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description fully discloses behaviors: creates workspace, returns apiKey, URL expires in 1 hour, single-use, IP-rate-limited. It also describes the agent-first signup flow, ensuring the agent understands the side effects and constraints.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single paragraph that front-loads the trigger condition, then covers main action, preferred flow, and fallback. It is concise but dense; could be more structured, yet every sentence adds value.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness5/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given 3 optional parameters, no output schema, and the signup context, the description covers purpose, trigger, flow, return values (apiKey, deliveryUrl), and constraints. It is complete for an agent to select and invoke without gaps.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100%, so baseline is 3. The description adds value by explaining that agentName logs for support, contactEmail for future notifications, and workspaceLabel defaults to 'default'. This provides context beyond schema descriptions.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool creates a new workspace and returns an API key, triggered by a 401 error when the user has no API key. This is a specific verb+resource and distinguishes from sibling tools like chieflab_brain_read.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines5/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Explicitly says 'USE WHEN' with the condition (no API key and 401 error). Provides a preferred flow (write to MCP config) and a fallback (deliver URL), giving clear guidance on when and how to use.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_use_manual_fallbackAInspect

USE WHEN a launch action is blocked because the channel has no automated publish path (Product Hunt, Hacker News, Reddit, Discord, Slack communities). Returns a paste-ready manual-publish brief: the exact text to post, a per-platform checklist (best time, tags/flair, who to ping, first-comment script), the URL the user pastes back to ChiefLab once posted, and a measurement template the agent can call 24h later via chiefmo_record_proof. Converts a 'blocked' channel into a 5-minute human-in-the-loop ship. Approval-gated upstream: the originating publishAction must already be 'approved' on the reviewUrl before this tool returns the brief.

ParametersJSON Schema
NameRequiredDescriptionDefault
channelNoOverride the channel inferred from the action. One of: product_hunt | hacker_news | reddit | discord | slack | indie_hackers | dev_to | ycombinator.
actionIdYesThe publishAction.id from chiefmo_launch_product whose connector is 'manual_only'. ChiefLab pulls the brief, channel hints, and brand voice from this action's preflight.
subredditNoOptional. For channel=reddit, which subreddit (e.g. 'r/SaaS', 'r/AI_Agents'). ChiefLab's brief picks the right post format per sub.
workspaceIdNoOptional workspace id.
discordCommunityNoOptional. For channel=discord, which community (the brief tailors tone + announcement-channel pick).
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations provided. Description mentions approval requirement and returns a brief with tracking template, but does not disclose side effects, permission needs, or whether it modifies state.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Three sentences, front-loaded with usage condition and output. Could be slightly trimmed but is efficient and well-structured.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given no output schema, description explains return components (text, checklist, URL, measurement template) and integrates with other tools (chiefmo_record_proof). Provides sufficient context for agent use.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema covers 100% of parameters. Description adds qualitative context (e.g., 'Override the channel', 'ChiefLab pulls the brief, channel hints, and brand voice'). Adds value beyond schema.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Description clearly states the tool is for blocked manual-only channels, returns a paste-ready brief. Differentiates from siblings like chiefmo_publish_approved_post (automated) and chieflab_launch_product (general launch).

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Explicitly says 'USE WHEN a launch action is blocked because the channel has no automated publish path' and 'Approval-gated upstream.' Provides clear context but does not explicitly list alternatives or when not to use.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_workspace_membersAInspect

List workspace members with roles (owner / approver / editor / viewer). Use for team management and to see who can approve publishes.

ParametersJSON Schema
NameRequiredDescriptionDefault
workspaceIdNo
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description must convey behavioral traits. It correctly identifies the tool as a read/list operation ('List workspace members'), but does not disclose any potential limits, pagination, or authorization requirements. The description is adequate for a simple read tool.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, focused sentence that conveys the tool's purpose and a common use case. It is not overly verbose, but could include a brief note on the parameter.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

The description covers the core functionality and a use case, but lacks details on required input (workspaceId) and return format. Given the tool's simplicity (one param, no output schema), the description meets minimum viability but is not fully comprehensive.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters2/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The schema has one parameter (workspaceId) with 0% description coverage. The description does not mention the parameter or explain that it is required to filter by workspace. The agent must infer the parameter from context, which reduces clarity.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states 'List workspace members with roles' and enumerates specific roles (owner/approver/editor/viewer), making the tool's purpose very specific. It distinguishes itself from sibling tools like chieflab_set_workspace_member.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides a use case: 'Use for team management and to see who can approve publishes.' This gives context for when to use it, but does not explicitly mention when not to use or alternatives.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chieflab_workspace_metricsAInspect

Get scale metrics for a workspace — launches, repeat-launch rate, approval rate, publish rate, measured-outcome rate, open blockers, cadence breakdown (24h/3d/7d). Use for the scale dashboard or 'how is this account doing' checks.

ParametersJSON Schema
NameRequiredDescriptionDefault
workspaceIdNoOptional. Defaults to your API key's workspace.
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations, so description carries full burden. Describes a read operation but adds no behavioral details beyond purpose—no mention of side effects, rate limits, or response size. Adequate but minimal.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two efficient sentences: first defines action and metrics, second gives usage guidance. No wasted words, front-loaded with purpose.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Lists all metrics but lacks output format description (e.g., JSON structure). Given low complexity and one optional parameter, this is nearly complete.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema covers 100% of parameters; description adds no new meaning. The optional workspaceId is already described in schema. Baseline score of 3.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the action ('Get scale metrics') and lists specific metrics (launches, repeat-launch rate, etc.), distinguishing it from sibling tools like chiefmo_measure_launch_results which likely focus on other metrics.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Explicitly says 'Use for the scale dashboard or 'how is this account doing' checks', providing clear context. However, it does not mention when not to use or offer alternatives.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_approve_actionAInspect

Approve one ChiefMO publish/send action so its executor (publish_approved_post / send_approved_email) can fire. USE WHEN the user — in IDE chat — said 'approve ' (e.g. 'approve linkedin', 'approve hn'), 'approve all', 'ship it', 'go ahead', or otherwise greenlit a specific draft you rendered. Match the user's words to the channel, look up agentGuide.renderInChat[channel].actionId from the launch response, and call this tool with that actionId. This is the IDE-native approval path — no need to push the user to the reviewUrl. Pass actionId (preferred) or id (legacy alias).

ParametersJSON Schema
NameRequiredDescriptionDefault
idNoLegacy alias for actionId.
actionIdNoThe action id from publishActions[].id in a launch_product response. Preferred.
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations provided, so description must cover behavioral traits. It discloses the approval effect and executor activation but does not mention error handling, idempotency, or side effects of approving an invalid actionId.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Description is concise, front-loading the core purpose in the first sentence. It efficiently conveys usage context without extraneous detail. Could be slightly tightened but overall well-structured.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given no annotations or output schema, the description covers the high-level workflow, usage trigger, and parameter derivation. Sufficient for an agent to select and invoke the tool correctly, though edge cases (invalid actionId) are not addressed.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100% with descriptions. Description adds context: actionId is preferred, id is legacy alias, and explains how to derive the value from launch response. This enhances schema information.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool approves a specific ChiefMO publish/send action, enabling its executor to fire. It specifies the verb 'approve' and the resource (action), distinguishing it from sibling executors like publish_approved_post.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Explicitly tells when to use: when user says 'approve <channel>', 'approve all', etc., in IDE chat. Provides detailed workflow on how to obtain the actionId from the launch response. Lacks explicit 'when not to use' but is otherwise strong.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_approve_runCInspect

Approve a ChiefMO run that is awaiting approval.

ParametersJSON Schema
NameRequiredDescriptionDefault
idYes
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

The description says 'Approve' which implies a state change, but no annotations are provided. It does not disclose consequences, required permissions, reversibility, or side effects of approval.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness3/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single concise sentence, but it is so brief that it sacrifices informativeness. It could include more details without becoming verbose.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool has one parameter and no output schema, the description should explain the parameter and the outcome of approval. It fails to provide this context, leaving the agent with minimal information to use the tool correctly.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters2/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The only parameter 'id' has no description in the schema (0% coverage) and the tool description does not explain what the id represents, such as whether it's a run ID or another identifier. The format or source of the id is missing.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the action (approve) and the object (ChiefMO run that is awaiting approval). However, it does not differentiate from the similar sibling tool 'chiefmo_gtm_run_approve' or explain the relationship to 'chiefmo_reject_run'.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance is provided on when to use this tool versus alternatives like 'chiefmo_gtm_run_approve' or under what circumstances a run should be approved. The description lacks context on prerequisites or conditions.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_check_claimsAInspect

Scan a draft (post / email / blog / ad / landing copy) for unsupported claims BEFORE publish. Each concrete claim is checked against the company brain (repo facts, proof assets, brand voice). Returns ranked list with status (verified | unsupported | risky) and suggested fixes. Run as a publish gate.

ParametersJSON Schema
NameRequiredDescriptionDefault
goalYesThe draft text to scan, or a description of what to check.
claimsNoOptional pre-extracted list of claims. If omitted, the skill extracts them from the draft.
brandUrlNoOptional brand URL.
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description carries full burden. It accurately describes the check behavior and return format but omits any discussion of side effects, rate limits, or authentication needs. No contradictions exist.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, well-structured paragraph of three sentences. It front-loads the core action and quickly adds context and output details. No extraneous information.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness5/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given no output schema, the description explains the return format (ranked list with status and suggested fixes). It also notes dependencies (company brain) and is complete for a pre-publish gate tool.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100% but description adds value by explaining that claims are checked against the 'company brain (repo facts, proof assets, brand voice)'. This gives meaning beyond the schema's parameter names.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Description clearly identifies the tool as scanning drafts for unsupported claims before publishing. It specifies content types (post, email, blog, ad, landing copy) and expected output (ranked list with statuses and fixes). This differentiates it from siblings like chiefmo_draft_blog_post or chiefmo_generate_ad_variants.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description explicitly states 'Run as a publish gate' and 'BEFORE publish', providing clear context for when to use. However, it does not mention when not to use or list alternative tools, leaving some ambiguity for the agent.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_company_brainAInspect

Return the full rendered company brain for a workspace (brand voice + repo facts + personas + proofs + channel performance + competitor watch + voice samples). Use to inspect what ChiefLab knows, debug grounding, or pipe into a custom prompt.

ParametersJSON Schema
NameRequiredDescriptionDefault
formatNomarkdown (default, prompt-ready) | json (structured)
tenantIdNo
workspaceIdNo
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description carries the full burden. It implies a read-only operation ('Return') but does not explicitly state side-effects, authentication needs, or rate limits. The listing of output components adds some transparency, but more detail on behavior would be beneficial.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is two sentences, no fluff. First sentence defines output, second lists use cases. Front-loaded and efficient.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given no output schema and multiple output components, the description covers the main elements and use cases. It does not detail the structure of the returned data beyond listing sections, which is sufficient for an agent to understand the tool's capability.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is only 33% (only 'format' described). The description clarifies that 'format' defaults to markdown and accepts 'json', adding value. However, 'tenantId' and 'workspaceId' are left without explanation, so the description only partially compensates for the low schema coverage.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool returns the 'full rendered company brain' and lists its components. The verb 'Return' and resource definition are specific. However, it does not explicitly differentiate from the sibling 'chieflab_brain_read', which may serve a similar function, though the description is sufficiently detailed.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description explicitly lists use cases: 'inspect what ChiefLab knows, debug grounding, or pipe into a custom prompt.' This provides clear context for when to use the tool, but it does not mention when not to use it or name alternatives, which would improve the score.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_continue_launch_loopAInspect

Resume a ChiefLab launch loop from runId. USE WHEN an agent has already called chieflab_get_users_after_build / chiefmo_launch_product and needs the exact next action: surface reviewUrl, execute an approved action, wait for measurement, measure results, or prepare the next move. This exists because agents are stateless; ChiefLab remembers approval state, actionIds, connector blockers, measurement windows, and next tool calls.

ParametersJSON Schema
NameRequiredDescriptionDefault
runIdYesThe launchId/runId returned by chieflab_get_users_after_build or chiefmo_launch_product.
workspaceIdNoOptional workspace id. Hosted auth usually supplies this.
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description clearly explains the tool's behavior: it can surface reviewUrl, execute an approved action, wait for measurement, measure results, or prepare the next move. It also notes that ChiefLab remembers approval state, actionIds, etc., providing good behavioral context.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is two sentences, front-loaded with the primary purpose, and uses clear, direct language. Every part earns its place, with no redundancy or extraneous information.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's complexity (state-resuming loop with multiple possible actions) and the absence of an output schema, the description provides sufficient context about what the tool does and its place in the workflow. It could mention potential return values more explicitly, but overall it is complete.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The input schema has 100% coverage, so baseline is 3. The description adds that runId is the launchId/runId returned by prior tools and that workspaceId is optional (usually supplied by auth), which adds useful context beyond the schema.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states it resumes a ChiefLab launch loop from a runId, specifying the verb 'Resume' and the resource 'launch loop'. It also distinguishes itself from siblings by referencing the context of prior calls like chieflab_get_users_after_build or chiefmo_launch_product.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines5/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description explicitly states when to use this tool ('USE WHEN an agent has already called chieflab_get_users_after_build / chiefmo_launch_product and needs the exact next action') and explains the stateful behavior of ChiefLab, which helps the agent understand why this tool exists and when it is appropriate.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_create_runBInspect

Create a ChiefMO marketing run from a business goal.

ParametersJSON Schema
NameRequiredDescriptionDefault
goalYesThe marketing outcome the user wants.
workspaceIdNoOptional workspace id. Defaults to local.
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations provided, and the description does not disclose any behavioral traits, side effects, permissions, or what 'create' entails beyond the minimal description.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single concise sentence with no extraneous information, but front-loading is limited.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given no output schema and no annotations, the description is too minimal. It does not explain what the tool returns or what side effects occur, leaving significant gaps for a creation action.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100%, so the description does not add significant meaning beyond the parameter descriptions. Baseline 3 is appropriate.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool creates a ChiefMO marketing run from a business goal, using a specific verb and resource. It is distinct from sibling tools like chiefmo_gtm_run_start.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance on when to use this tool versus alternatives, such as chiefmo_gtm_run_start or other creation tools. Missing context on prerequisites or scenarios.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_create_social_graphicsAInspect

Generate on-brand social/launch graphics via Gemini 2.5 Flash Image. USE WHEN the user asks for launch images, social graphics, hero images, or 'visuals for the launch.' Note: chiefmo_launch_product already includes graphics inline — only call this directly for ad-hoc image generation outside a launch run. Metered separately (~$0.04 cost / 50 retail credits per image).

ParametersJSON Schema
NameRequiredDescriptionDefault
brandNoOptional inline brand context (palette, type, style).
countNoHow many to generate (default 1, max 4).
promptYesImage prompt. Brand context auto-applied if tenantId or brand is passed.
tenantIdNoOptional. Loads stored brand context for grounding.
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations provided, so description carries full burden. It discloses the AI model (Gemini 2.5 Flash Image), that brand context is auto-applied, and metering details with cost and credits. Does not mention auth or destructive behavior, but for a generation tool this is adequate.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Three concise sentences: first states purpose and model, second gives usage guidance, third adds sibling distinction and cost. No redundant information.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a graphics generation tool, the description covers what it does, when to use, cost, and relationship to sibling. Missing output format, but this is typical for gen tools. Overall complete for its complexity level.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100%, so baseline is 3. Description adds minor context about brand auto-application but otherwise does not extend meaning beyond what the input schema already provides.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool generates 'social/launch graphics' and lists specific use cases: launch images, social graphics, hero images, visuals for launch. It distinguishes from sibling tool chiefmo_launch_product by noting that tool includes graphics inline, so this one is for ad-hoc generation.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines5/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Explicit 'USE WHEN' directive lists scenarios (launch images, social graphics, etc.). Also explicitly states when NOT to use: if part of a launch run, as chiefmo_launch_product already handles it. Provides alternative by name.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_design_experimentAInspect

Convert a marketing question into a runnable experiment: hypothesis, variants, primary success metric, sample-size guidance, and kill criteria. Use BEFORE shipping a new test so you have a stop rule. Manual mode is the entire spec; with GA4/Search Console connected, baselines are pulled live.

ParametersJSON Schema
NameRequiredDescriptionDefault
goalYesThe marketing question or change you want to test.
brandUrlNoOptional brand URL.
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description carries the full burden. It discloses that baselines are pulled live when GA4/Search Console are connected, but it does not mention any side effects, required permissions, or error handling. For a design tool that likely has no destructive effects, this is adequate but not thorough.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is two sentences with no unnecessary words. The first sentence defines the core function, and the second adds usage context. Every sentence earns its place, making it highly concise and well-structured.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given no output schema, the description lists the key components of the output (hypothesis, variants, etc.), which is sufficient for an agent to understand the return value. It could mention the output format or structure, but the listed components provide reasonable completeness for a design tool.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100%, so the parameters are already documented. The description does not add extra meaning beyond what the schema provides; it only implies that 'goal' is the marketing question. A 3 is appropriate when the schema already handles parameter description effectively.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool's purpose: converting a marketing question into a runnable experiment with specific components (hypothesis, variants, metric, sample size, kill criteria). It distinguishes itself from sibling tools like chiefmo_diagnose_anomaly or chiefmo_draft_blog_post by focusing on experiment design.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description explicitly says 'Use BEFORE shipping a new test so you have a stop rule', providing clear when-to-use guidance. It also describes two modes (manual vs. connected to GA4/Search Console), though it does not explicitly state when not to use or list direct alternatives.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_diagnose_anomalyAInspect

Analytics anomaly explainer. Given a metric drop or spike (e.g. 'Search Console clicks down 30%'), ranks likely causes and proposes safe corrective actions.

ParametersJSON Schema
NameRequiredDescriptionDefault
goalYesThe metric anomaly to diagnose.
brandUrlNoOptional brand URL for context.
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations provided; the description implies read-only analysis by 'explainer' and 'proposes actions', but does not explicitly state safety or potential side effects. Could be more transparent about non-destructive nature.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two efficient sentences that front-load the purpose and provide an example. No wasted words.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Covers the tool's purpose, input example, and expected output (ranked causes and actions). Without an output schema, it briefly describes the return value, which is sufficient for this diagnostic tool.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100% so the schema fully documents parameters. The description aligns with the 'goal' parameter via the example, but adds no additional semantic nuance beyond the schema.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states it is an analytics anomaly explainer, gives a concrete example ('Search Console clicks down 30%'), and specifies that it ranks causes and proposes actions, distinguishing it from similar diagnostic tools.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

It explicitly says 'Given a metric drop or spike', indicating when to use it. However, it does not mention when not to use it or provide alternatives, but the context is clear enough for the intended use case.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_diagnose_marketingAInspect

Diagnose an EXISTING marketing program (not a new launch — use chiefmo_launch_product for that). USE WHEN the user already has live campaigns and asks 'why is X not working' or 'iterate on this funnel.' Returns evidence, drafted assets, proposed actions, signed reviewUrl, per-tenant memory. Approval-gated. For first launches, prefer chiefmo_launch_product.

ParametersJSON Schema
NameRequiredDescriptionDefault
goalYesThe marketing outcome, problem, or question to diagnose. e.g. 'Help this newly built SaaS get its first 100 users.'
brandNoOPTIONAL but recommended for cost + accuracy. If your agent already knows the brand context (you're inside Cursor with the codebase, you have profile data, you have prior conversation), pass it directly here and ChiefLab skips brand scanning entirely. Cheapest + most accurate path. Omit only if you want ChiefLab to derive brand context from tenantId or tenantUrl.
tenantIdNoOptional. Identifies the developer's end-user. When set, ChiefMO loads per-tenant brand context, voice samples, and memory. Use chieflab_create_tenant + chieflab_set_tenant_context first to establish context.
tenantUrlNoOptional shortcut. If you don't have a tenantId yet, pass the end-user's product URL and ChiefMO will extract brand context on the fly (does not persist — use chieflab_set_tenant_context for that).
outputModeNoHow ChiefLab should produce skill outputs. "context" (default, cheapest): returns drafting briefs your LLM renders into final copy — ChiefLab does brand context, evidence, prompts, and approval rails, your agent's LLM does the writing. "draft": ChiefLab generates rough drafts using cheap models. "full": ChiefLab generates final copy server-side (premium credits). Recommended: leave default unless you specifically need server-side generation.
webhookUrlNoOptional HTTPS URL ChiefLab POSTs to when run completes (signed with X-ChiefLab-Signature).
workspaceIdNoOptional workspace id. Defaults to the workspace derived from the API key.
idempotencyKeyNoOptional dedupe key (24h cache). Same key returns cached result instead of re-running.
includeSnapshotsNoWhether the caller expects connector snapshot evidence when available.
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations exist, so the description carries the full burden. It mentions that the tool is approval-gated and returns specific outputs (evidence, drafted assets, proposed actions, reviewUrl, per-tenant memory). It does not disclose any destructive potential or rate limits, but as a diagnostic tool, these are less critical. The description is transparent about its behavior and return value.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single paragraph but front-loads key information: purpose and when to use. It is concise and each sentence adds value. However, it could be slightly more structured (e.g., bullet points for return types) for easier parsing, but it remains efficient.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's complexity (9 parameters, nested objects, no output schema), the description covers the essential aspects: purpose, usage conditions, parameter guidance, return types, and the approval gate. It includes optional parameters like webhookUrl and idempotencyKey. It is sufficiently complete for an agent to use the tool correctly.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100% (all 9 parameters are described). The description adds value beyond the schema by explaining the brand object's usage: 'OPTIONAL but recommended for cost + accuracy' and conditions for omitting. It also explains the outputMode parameter options in detail, aiding the agent in choosing the right mode.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states it diagnoses an EXISTING marketing program, specifies it is not for new launches, and contrasts with the sibling 'chiefmo_launch_product'. It lists what it returns (evidence, drafted assets, proposed actions, etc.), making the tool's purpose unambiguous.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines5/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description explicitly tells when to use this tool: when the user has live campaigns and asks 'why is X not working' or wants to 'iterate on this funnel.' It also explicitly tells when not to use it: 'For first launches, prefer chiefmo_launch_product.' This provides clear decision criteria.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_draft_blog_postAInspect

Draft a full repo-aware blog post (title, slug, meta description, body, internal links, OG image prompt). Manual mode emits markdown you paste into your CMS; with GitHub connected, ChiefLab can commit the post directly. With Search Console / GA4 connected, the post is tracked after publish.

ParametersJSON Schema
NameRequiredDescriptionDefault
goalYesWhat this post should accomplish or what it's about.
brandUrlNoOptional brand URL for context.
repoContextNoOptional repo context — increases grounding so the post references real product details, not invented ones.
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description carries full burden. It discloses that the tool can either output markdown for manual pasting or commit directly to GitHub if connected, and that post tracking is available with Search Console/GA4. It does not explicitly mention potential side effects like file creation or overwrite, but the commit behavior is clearly indicated.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two efficient sentences covering all key aspects (components, modes, tracking). No wasted words, front-loaded with the primary action.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given no output schema, the description explains what is produced (markdown or committed post) and the components. It covers integrations and the need for prior connections. Minor gap: does not detail the return format or confirmation message, but the output is sufficiently described for an agent.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, so baseline is 3. The description does not add significant meaning beyond the schema: it repeats the integration context but does not elaborate on parameter usage or syntax.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Description clearly states the tool drafts a blog post with specific components (title, slug, meta description, body, internal links, OG image prompt) and explains two modes (manual markdown vs. GitHub commit). It distinguishes from sibling content tools by mentioning 'repo-aware' and the integration with GitHub and Search Console/GA4.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Description explains the modes and integrations but does not explicitly state when to use this tool over alternatives like `chiefmo_generate_landing_copy` or `chiefmo_generate_social_posts`. Usage is implied but no exclusions or comparisons are provided.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_extract_brand_dnaAInspect

USE WHEN the user wants to understand a brand (theirs or a competitor) — 'analyze brand X', 'what is X's positioning', 'extract brand voice from this URL', 'figure out X's content pillars'. Brand DNA discovery: given a URL or brand name, returns category, audience, voice, strategy pillars, competitors, visual cues, content angles, and risks. Use as the first call when you have a brand to work on — every other ChiefMO skill uses the result as context.

ParametersJSON Schema
NameRequiredDescriptionDefault
brandUrlYesURL or brand name (e.g. 'https://stripe.com' or 'Stripe').
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description carries full responsibility. It lists outputs but does not disclose behavioral traits such as whether it is read-only, required permissions, or side effects. Without annotations, more detail would be helpful.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two sentences front-load purpose with 'USE WHEN' and examples. Second sentence outputs key information. No extraneous text, highly efficient.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness5/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Despite the absence of an output schema, the description lists all key output categories (category, audience, voice, etc.). It also notes the tool's role as a foundational step for other skills, providing complete context for an agent.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100% for the single parameter 'brandUrl', and the description adds 'URL or brand name' which mirrors the schema description. No additional semantic meaning beyond what the schema provides.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description uses specific verb phrases like 'analyze brand X' and 'extract brand voice from this URL', clearly identifying the tool's purpose of understanding a brand. It distinguishes from siblings by positioning it as the first call for brand work, which other ChiefMO skills depend on.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines5/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Explicitly starts with 'USE WHEN' and provides concrete examples of user requests. It also states it should be used as the first call and that other skills use its result as context, giving clear usage context and dependencies.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_extract_repo_factsBInspect

Pull structured product facts from the supplied repoContext (commits, changed files, routes, README, pricing) and emit a list ready to persist.

ParametersJSON Schema
NameRequiredDescriptionDefault
goalYesOptional framing (e.g. 'features only', 'pricing changes').
brandUrlNo
repoContextNoSame shape as chiefmo_launch_product.repoContext.
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations provided. The description says 'emit a list ready to persist' but doesn't clarify if the tool itself persists data or is read-only. This ambiguity leaves the agent unsure of side effects.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Single sentence of 20 words, front-loaded with the verb 'Pull'. No unnecessary words.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Despite having a nested repoContext parameter and no output schema, the description provides minimal detail about input constraints or the output format. Essential context is missing.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters2/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 67% with brandUrl undocumented. The description adds no parameter context beyond what the schema provides, failing to clarify the missing brandUrl parameter.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool extracts structured product facts from repoContext and outputs a list for persistence. It distinguishes from siblings like chiefmo_record_repo_fact which records individual facts.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies use when repoContext is available and product facts are needed, but no explicit when-not or alternatives are mentioned. Siblings like chiefmo_record_repo_fact are not referenced.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_generate_ad_variantsAInspect

USE WHEN the user wants paid-ad copy — 'write Google Ads variants', 'Meta ad creative for this campaign', 'split-test 5 headlines for X'. Returns ad variants (angles, headlines, primary text, CTAs) plus a small test plan. For Meta and Google Ads. Drafts only — publishing requires approval and budget rules.

ParametersJSON Schema
NameRequiredDescriptionDefault
goalYesAd campaign objective and target audience.
brandUrlNoOptional brand URL for context.
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries full burden. It discloses the output (ad variants with angles, headlines, etc., plus a test plan), the draft-only nature, and platform scope. It does not cover authentication, rate limits, or error conditions, but for a generative tool the disclosed behaviors are sufficient and accurate.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is two sentences, front-loaded with the usage trigger, then concise output and constraint details. No redundant or extraneous information.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a simple generative tool (2 params, no output schema, no annotations), the description provides key context: what the tool does, what it returns, and important constraints (drafts only, platforms). It could elaborate on the test plan or variant count, but overall covers the essentials.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100% (both parameters described in schema). The description does not add extra detail beyond the schema: it implies the 'goal' parameter captures the ad objective and target audience, but no syntax or format guidance. Baseline 3 is appropriate.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool generates paid-ad copy for Meta and Google Ads, with explicit user-intent examples like 'write Google Ads variants' and 'Meta ad creative for this campaign'. It distinguishes from sibling tools (e.g., generate_landing_copy, draft_blog_post) by specifying the paid-ad context and output (variants plus test plan).

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description begins with 'USE WHEN' and provides concrete triggers ('write Google Ads variants', 'split-test headlines'). It also states platforms ('For Meta and Google Ads') and a key limitation ('Drafts only — publishing requires approval and budget rules'). However, it does not explicitly name alternative tools for other copy types, leaving some inference needed.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_generate_email_sequenceAInspect

USE WHEN the user wants an email-only sequence (not a full launch) — 'draft a welcome sequence', 'win-back emails for churned users', 'nurture series for new signups'. Returns 4-email skeleton with subject line ladder, send timing (Day 0 / 2 / 5 / 9), and segmentation notes. Drafts only — sending requires approval and a connected ESP. For launch emails specifically, use chiefmo_launch_product (it bundles email + social + landing in one approval flow).

ParametersJSON Schema
NameRequiredDescriptionDefault
goalYesSequence purpose (welcome, nurture, win-back, launch, etc.).
brandUrlNoOptional brand URL for context.
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries full burden. It states the tool returns a draft skeleton, requires approval and ESP connection, and implies creation behavior. It does not explicitly state whether it saves data or is idempotent, but overall discloses key behavioral traits.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, well-structured paragraph of three sentences. It is front-loaded with the usage trigger condition and every sentence adds essential information without redundancy.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness5/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a tool with 2 simple parameters and no output schema, the description fully explains the return value (4-email skeleton with specific details) and necessary prerequisites (approval, ESP). No additional information is needed.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The input schema has 100% coverage with descriptions for both parameters (goal and brandUrl). The description adds value by providing example values for goal (welcome, nurture, win-back, launch, etc.) and clarifying brandUrl as optional context.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool generates an email-only sequence (not a full launch) and specifies the output: a 4-email skeleton with subject line ladder, send timing, and segmentation notes. It distinguishes itself from the sibling tool chiefmo_launch_product.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines5/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description starts with 'USE WHEN' and provides concrete examples (welcome, win-back, nurture). It explicitly says what the tool is not for ('not a full launch') and directs the agent to use chiefmo_launch_product for launch emails. It also notes that the tool only drafts, requiring approval and a connected ESP.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_generate_landing_copyAInspect

USE WHEN the user wants landing-page copy — 'write a landing page for X', 'rewrite my hero', 'give me value props for this product', 'FAQ for the pricing page'. Returns ready-to-paste markdown: hero headline, subhead, value props, social proof slots, FAQ, CTA. Page build/publish requires approval and a connected web tool.

ParametersJSON Schema
NameRequiredDescriptionDefault
goalYesWhat the landing page should sell or convert.
brandUrlNoOptional brand URL for context.
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description carries the full burden. It discloses output format ('ready-to-paste markdown') and a prerequisite ('Page build/publish requires approval and a connected web tool'), but does not cover other behavioral aspects like mutability, auth needs, or rate limits.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Three sentences with no wasted words. The first sentence efficiently combines usage guidelines with examples, the second describes output, and the third adds a key prerequisite. Could be slightly more structured, but overall concise and front-loaded.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given no output schema, the description adequately explains the return format and includes a prerequisite. For a tool with two simple parameters, this covers the essential context. However, deeper details about markdown sections could improve completeness.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100% with both parameters described. The description adds context via examples but does not provide new semantic detail beyond the schema definitions. Baseline score of 3 is appropriate.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool's purpose with specific examples like 'write a landing page for X' and 'rewrite my hero'. It distinguishes itself from other content generation tools by focusing on landing-page copy, though it does not explicitly name siblings.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The 'USE WHEN' pattern provides clear context for when to use the tool, with explicit examples. However, it lacks any guidance on when not to use it or alternatives, leaving the agent to infer exclusion criteria.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_generate_social_postsAInspect

USE WHEN the user wants social-only drafts (no full launch) — 'draft an X post', 'write LinkedIn carousels for this', 'give me 5 Instagram captions about X'. NOT for new product launches (use chiefmo_launch_product instead — that returns the same plus on-brand graphics, approval gate, and per-channel briefs). Returns posts with platform-native hooks, captions, hashtags, and posting cadence. Drafts only; scheduling/publishing requires approval via chiefmo_publish_approved_post.

ParametersJSON Schema
NameRequiredDescriptionDefault
goalYesWhat the social posts should accomplish.
brandUrlNoOptional brand URL for context (will trigger Brand DNA discovery).
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

The description discloses what the tool returns (posts with hooks, captions, hashtags, cadence) and its draft-only nature, implying no side effects or destructive actions. Without annotations, this adds value beyond the schema.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, well-structured paragraph with no wasted words. It covers usage, exclusions, output, and lifecycle in a logical order.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness5/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given no output schema, the description fully explains the return format (posts with hooks, captions, hashtags, cadence) and the overall workflow (drafts only, needs approval). No gaps.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100%, so the description adds context: 'goal' is what posts accomplish, 'brandUrl' triggers Brand DNA discovery. This clarifies usage beyond the schema.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool's purpose: generating social-only drafts. It uses specific verbs ('draft', 'write', 'give me') and distinguishes from a key sibling ('chiefmo_launch_product') that handles full launches.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines5/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Explicit 'USE WHEN' conditions are provided, along with clear exclusion ('NOT for new product launches') and a named alternative ('chiefmo_launch_product'). It also notes that scheduling/publishing requires approval via 'chiefmo_publish_approved_post'.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_get_actionAInspect

Get one ChiefMO action by id.

ParametersJSON Schema
NameRequiredDescriptionDefault
idYes
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations exist, so description is the sole source for behavioral traits. It implies a simple read operation but does not explicitly state absence of side effects, authorization needs, or other behaviors. Adequate for a basic getter.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single concise sentence that immediately conveys the purpose. Every word earns its place; no fluff.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a simple fetch operation, the description is largely complete. However, it does not mention what the return value represents (e.g., full action object), though this may be inferred from context. Still sufficient for a single-parameter tool with clear siblings.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters2/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

With 0% schema description coverage, the description adds no meaning to the 'id' parameter beyond the schema's type string. It does not clarify expected format, examples, or constraints.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the verb 'Get' and the resource 'one ChiefMO action by id', distinguishing it from siblings like chiefmo_list_actions which lists multiple actions.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance provided on when to use this tool versus alternatives such as chiefmo_list_actions or chiefmo_get_run. The description lacks context for tool selection.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_get_assetBInspect

Get one ChiefMO generated asset by id.

ParametersJSON Schema
NameRequiredDescriptionDefault
idYes
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description carries the full burden. It only states 'Get one... by id' without disclosing any behavioral traits such as error handling, authentication requirements, or whether the operation is read-only (though implied). Minimal transparency.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is extremely concise: one short sentence with no redundancy. Every word serves a purpose.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the low complexity (single required param, read operation), the description is adequate but lacks details about the return value, error conditions, or expected format. No output schema exists, so the description should provide some context, which it does not.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters2/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The schema description coverage is 0%, and the description only repeats 'by id' which adds no additional meaning beyond the parameter name. It does not describe what the 'id' represents, how to obtain it, or any constraints (e.g., format).

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the verb 'Get' and the resource 'asset', and specifies filtering by 'id'. This distinguishes it from sibling tools like 'chiefmo_list_assets' which lists assets without filtering by a single ID.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage when you have a specific asset ID, but provides no explicit guidance on when to use this tool versus alternatives such as 'chiefmo_list_assets' or other get tools. No exclusions or context are given.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_get_runBInspect

Get one ChiefMO run by id.

ParametersJSON Schema
NameRequiredDescriptionDefault
idYes
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations provided. Description does not disclose if this is a read-only operation, what happens if the ID is invalid, or any permissions required. For a simple get, minimal disclosure but still lacking.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Single concise sentence with no unnecessary words. Efficiently communicates the core function.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the low complexity (1 param, no output schema, no nested objects), the description is minimally adequate. However, it leaves out return value details and error handling, which are important for an agent deciding whether to use this tool.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters2/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema has 0% description coverage for the single parameter 'id'. Description only says 'by id', adding minimal meaning beyond the schema. Should explain what id refers to or format expectations.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Description clearly states verb 'Get', resource 'ChiefMO run', and method 'by id'. Distinguishes from sibling tools like chiefmo_get_action and chiefmo_get_asset.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance on when to use this tool vs alternatives such as chiefmo_list_runs, chiefmo_approve_run, or chiefmo_get_skill. No context or exclusions provided.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_gtm_agent_runAInspect

Advanced. Run one specific GTM agent in isolation, bypassing the orchestrator. Use when you've already collected the upstream handoffs and want to invoke a single agent for re-runs / diffs.

ParametersJSON Schema
NameRequiredDescriptionDefault
inputsNoOptional explicit inputs for the agent (otherwise pulled from the run's prior handoffs).
agentIdYesOne of: launch, social, email, blog-seo, creative, analytics, experiment, approval
gtmRunIdYes
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, yet the description omits behavioral details such as side effects, idempotency, or whether it modifies state. The burden is on the description, but it only states action without transparency.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two sentences precisely deliver purpose and usage guidance with no wasted words. Front-loaded with the core action.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a tool with no output schema and nested objects, the description lacks details on return values, error conditions, or blocking/async behavior. It covers purpose and usage but not output or execution context.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 67%, and the description adds no new information about parameters beyond what the schema provides. The description's high-level usage context doesn't deepen parameter understanding.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Explicitly states 'Run one specific GTM agent in isolation, bypassing the orchestrator,' with a clear verb and resource. It distinguishes from sibling tools like chiefmo_gtm_run_start by mentioning isolation and re-runs/diffs.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Provides a concrete condition: 'use when you've already collected the upstream handoffs and want to invoke a single agent for re-runs / diffs.' This implies when not to use, though explicit alternatives are not named.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_gtm_handoffs_listAInspect

Read every handoff emitted in a GTM run, in order. Useful for the command center timeline and for debugging why an agent skipped (no input handoff found).

ParametersJSON Schema
NameRequiredDescriptionDefault
runIdYes
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description carries the burden. It implies a read-only operation by stating 'Read every handoff...' but does not explicitly mention safety, permissions, or lack of side effects. This is adequate but not thorough.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two concise sentences. The first states the purpose, the second provides usage context. No wasted words. Front-loaded with key information.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's simplicity (1 parameter, no output schema, no annotations), the description covers purpose and usage. However, it lacks parameter explanation and does not describe the return format. It is nearly adequate but has gaps.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters1/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The only parameter 'runId' has no description in the schema (0% coverage). The description does not explain what 'runId' is or how to format it. With 0% coverage, the description should compensate but fails to add any meaning.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Description clearly states 'Read every handoff emitted in a GTM run, in order.' The verb 'Read' and resource 'handoffs emitted in a GTM run' are specific. It distinguishes from sibling tools like chiefmo_gtm_run_get which deals with runs, not handoffs.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Description explicitly says 'Useful for the command center timeline and for debugging why an agent skipped (no input handoff found).' This provides concrete use cases. However, it does not mention when not to use or alternative tools, but the context is clear.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_gtm_memory_getAInspect

Read the shared memory for a workspace + tenant. Memory keys are agent-namespaced (e.g. 'audience.refinement', 'seo.targetQueries.recent'). Returns the full key-value snapshot.

ParametersJSON Schema
NameRequiredDescriptionDefault
keyNoOptional. If set, returns just that one key's value.
tenantIdNo
workspaceIdNo
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description bears the full burden. It states 'Read' implying no side effects, explains key namespacing, and return type. However, it doesn't disclose error handling or behavior for missing keys. Given simplicity, it's mostly transparent.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two sentences, front-loaded with core purpose. Every sentence adds value: first states action and scope, second gives key format and output. No redundancy or fluff.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Covers key purpose and output, but lacks details on mandatory scope parameters (tenantId, workspaceId) and behavior when key is omitted. No output schema, but description suffices for a simple read operation.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters2/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Only 33% schema description coverage; the description adds no value beyond the schema for 'key' and does not explain 'tenantId' or 'workspaceId' parameters. The purpose mentions workspace and tenant but doesn't map to parameters.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states it reads shared memory for a workspace and tenant, with examples of memory keys. It distinguishes itself from siblings like 'chiefmo_gtm_memory_set' (write) and 'chiefmo_memory_diff' (diff).

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No explicit guidelines on when to use this tool versus alternatives. While the purpose is clear, there is no mention of when to read vs. set or diff, or any preconditions. Implicitly, it's for memory retrieval, but lacking explicit direction.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_gtm_memory_setAInspect

Write a fact into shared memory. Use sparingly — agents normally write their own memory. Caller-asserted writes are useful for seeding context (e.g. 'audience.refinement = {...}' before a first run) or correcting an agent's mistake.

ParametersJSON Schema
NameRequiredDescriptionDefault
keyYes
valueYesJSON-serializable value.
tenantIdNo
workspaceIdNo
sourceAgentIdNoDefaults to 'caller'. Use an agent id if asserting on behalf of an agent.
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description carries full responsibility. It explains the caller-asserted nature, default sourceAgentId behavior, and appropriate use cases. However, it does not disclose whether writes are idempotent or if previous values are overwritten, but the description is still informative.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is two sentences, no redundancy, front-loaded with the core action, and the second sentence provides essential usage context. Every word adds value.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a tool with 5 parameters and no output schema, the description covers purpose and usage but lacks details on parameter semantics (key format, tenantId/workspaceId purpose) and does not mention return value or error handling. It is adequate but not fully complete.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 40%, only value and sourceAgentId have descriptions. The description adds the default for sourceAgentId but does not explain the other parameters (key, tenantId, workspaceId) beyond what the schema provides. Given low coverage, the description should compensate more.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description explicitly states 'Write a fact into shared memory,' which is a specific verb+resource. It further differentiates from agents writing their own memory, making the purpose clear and distinguishing from sibling tools.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines5/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description advises 'Use sparingly — agents normally write their own memory' and provides concrete scenarios ('seeding context' and 'correcting an agent's mistake'), which clearly guide when to invoke this tool versus relying on agent self-writing.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_gtm_run_approveAInspect

Approve all pending actions in a GTM run. Equivalent to clicking 'approve all' on the reviewUrl. Returns the actions that fired (publish, send) and any that were rejected by safety checks.

ParametersJSON Schema
NameRequiredDescriptionDefault
actorNoWho approved (email or agent id, for audit).
runIdYes
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, description carries full burden. It discloses return value (fired/rejected actions) but omits write implications, permissions, or reversibility. Adequate but not thorough.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two sentences, 25 words, no filler. Front-loads the action and intent. Every word earns its place.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given missing annotations and output schema, the description is sufficient for a simple approval but leaves gaps: no error cases, prerequisites, or response format. Adequate.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage 50% (actor described, runId not). Description adds no parameter info beyond schema. It implies runId is the run but doesn't explain format or necessity. Baseline moderate.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Clear verb 'approve' with specific resource 'all pending actions in a GTM run', effectively distinguishing it from siblings like approve_action (single action). The 'Equivalent to clicking approve all' analogy reinforces understanding.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Provides clear context for bulk approval but lacks explicit when-not-to-use or alternatives. The sibling list hints at single-action alternatives, but the description itself doesn't guide selection.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_gtm_run_getAInspect

Read a GTM run by id. Returns the run row, every agent run, every handoff. Use this in the human command center / dashboard or to inspect a run after the fact.

ParametersJSON Schema
NameRequiredDescriptionDefault
runIdYes
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Describes it as a read operation and states the return structure. No annotations, so description carries full burden. Could note idempotence but sufficient for a read tool.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two sentences, front-loaded with purpose, then returns and use case. No extraneous information.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Adequately covers purpose, returns, and usage context. Lacks response format details but acceptable for a simple read tool.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Only one parameter 'runId' with no schema description. Description mentions 'by id' implying its purpose, but adds little beyond the schema. Baseline 3 due to low schema coverage.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Specifically states 'Read a GTM run by id' and details what is returned (run row, agent runs, handoffs). Clearly distinguishes from siblings like list or start.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Explicitly tells when to use ('in the human command center / dashboard or to inspect a run after the fact'). Does not mention when not to use or alternatives, but context is clear.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_gtm_run_listCInspect

List recent GTM runs in the caller's workspace.

ParametersJSON Schema
NameRequiredDescriptionDefault
limitNoDefault 20, max 100
workspaceIdNo
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations exist. Description only says 'list recent GTM runs' but does not clarify ordering, pagination, or what 'recent' means. Behavioral traits are minimal.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Single sentence, no fluff. Could add more context without losing conciseness, but efficiently conveys core purpose.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

No output schema, few parameters, no annotations. Description fails to specify return format, default behavior for workspaceId, or ordering. Missing essential details for a list tool.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters2/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage 50% (only limit described). Description adds no extra meaning for workspaceId or limit beyond what schema says. Does not compensate for missing schema descriptions.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Description clearly states verb 'list' and resource 'GTM runs' with scope 'recent' and 'in caller's workspace'. Distinct from siblings like chiefmo_gtm_run_get or chiefmo_gtm_run_start.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance on when to use this vs alternatives like chiefmo_list_runs or other GTM tools. No exclusions or context provided.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_gtm_run_startAInspect

USE WHEN the user wants the full multi-agent GTM team coordinated on a goal — 'launch this and run social/email/blog/creative together', 'do the whole GTM motion', 'plan + draft + approve + measure end-to-end'. Spawns a GTM run; the orchestrator picks which of the 8 agents (launch / social / email / blog-seo / creative / analytics / experiment / approval) to activate based on the goal. Returns the run id, the run plan (which agents activated and why), and links to /app/gtm for the human command center.

ParametersJSON Schema
NameRequiredDescriptionDefault
goalYesWhat the user wants accomplished. e.g. 'Launch this product and get our first 100 users.'
tenantIdNoOptional tenant id for per-tenant context + memory.
productUrlNoOptional product URL — gives downstream agents something concrete to ground against.
forceAgentsNoAdvanced. Override the launch agent's runPlan with an explicit list of agent ids to activate. e.g. ['social','creative'] to run only those two.
repoContextNoOptional repo context the agent gathered (routes, recentCommits, README excerpt, package.json highlights). The launch agent uses this for repo-grounded positioning.
workspaceIdNoOptional workspace id (defaults to caller's workspace).
skipPostPublishNoIf true (default for human-driven runs), the orchestrator stops at awaiting_approval and waits for the user to approve on /app/gtm. If false (background/cron path), it continues straight through analytics + experiment without an approval pause.
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations provided, so description covers behavior. It explains spawning a run, orchestrator agent selection, return values (run id, plan, links), and advanced options like forceAgents and skipPostPublish. Could mention idempotency or side effects like creating a persistent run object.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Description is well-structured with a usage instruction first, then details. It is efficient but could be slightly tighter by removing redundant phrases like 'the orchestrator picks which...' which is implied.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given complexity (multi-agent orchestration, 7 params, nested objects, no output schema), description covers key behavioral aspects: what it does, returns, and parameter roles. Could mention that it's a long-running operation or requires specific permissions, but overall sufficient.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, so baseline is 3. The description adds context for forceAgents ('overrides') and goal ('orchestrator picks agents'), but most parameter meaning is already in the schema.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Description clearly states the tool orchestrates a multi-agent GTM run on a goal, distinguishing it from sibling tools that handle single agents or specific tasks. It specifies the verb ('spawns'), resource ('GTM run'), and scope ('full multi-agent GTM team').

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Explicitly states 'USE WHEN' for full GTM coordination. Provides context for when to use, but does not explicitly state when not to use or offer alternatives among siblings like chiefmo_gtm_agent_run for single-agent tasks.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_launch_healthAInspect

Inspect launch-loop health for a run. USE WHEN an agent or user asks 'what is blocked?', 'what approvals are open?', 'what can I execute?', 'is measurement due?', or 'what should happen next?' Returns open approvals, blocked channels, pending execution, measurement windows, and next moves.

ParametersJSON Schema
NameRequiredDescriptionDefault
runIdNoThe launchId/runId to inspect. Strongly recommended.
workspaceIdNoOptional workspace id. Hosted auth usually supplies this.
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations provided, so description carries full burden. It implies read-only ('inspect', 'returns') but does not explicitly state safety, side effects, or permissions. Adequate but not detailed.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two sentences with no fluff. Front-loads purpose and usage triggers. Every word adds value.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness5/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Despite no output schema, description lists return categories (open approvals, blocked channels, etc.), making the tool's output predictable. No missing context.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema covers 100% of parameters. Description adds value by noting 'Strongly recommended' for runId and clarifying workspaceId's optionality and typical source.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states 'Inspect launch-loop health for a run' and lists specific queries it answers, differentiating it from sibling tools like chiefmo_continue_launch_loop and chiefmo_get_run.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Explicitly says 'USE WHEN' with example queries, providing clear usage context. However, it does not mention when not to use or alternative tools, but the listed use cases are sufficient.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_launch_productAInspect

Repo-aware launch operator for agent-built products. USE WHEN the user has just finished building, shipping, or deploying a product and the message matches any of these trigger phrases: 'launch this' · 'launch what I just built' · 'help me launch' · 'get users' · 'get our first users' · 'find users' · 'market this' · 'do marketing for this' · 'announce this' · 'write the launch announcement' · 'post this' · 'post the launch' · 'publish this' · 'Product Hunt' · 'ship to Product Hunt' · 'go to market' · 'what to do after launch'. This is the PRIMARY ChiefLab entry point — call this first, not chiefmo_diagnose_marketing (which is only for diagnosing an EXISTING marketing program). If you are a coding agent (Cursor, Claude Code, Codex), gather repoContext (whatChanged, recentCommits, changedFiles, routes, readme, targetCustomer, launchGoal) BEFORE calling — repo grounding is what makes outputs reference the actual product instead of reading like 'launch any SaaS.' Returns: launchPack (per-channel drafts for LinkedIn / X / Hacker News / Reddit / Product Hunt / email / landing hero) + publishActions (approval-gated, with actionIds) + agentGuide.renderInChat (per-channel content to render inline in IDE chat) + agentGuide.nextToolCalls.primary.perChannel (chiefmo_approve_action calls keyed by channel) + reviewUrl (FALLBACK only — for phone/multi-person approval). IDE-NATIVE FLOW: render each channel's draft inline in chat, wait for user to say 'approve ' or 'approve all', call chiefmo_approve_action per approved action. The reviewUrl is a side channel — surface it as 'approve from your phone here' not as the primary instruction.

ParametersJSON Schema
NameRequiredDescriptionDefault
icpNoOptional ICP block. Overrides repoContext.targetCustomer + brand.audience for ICP grounding.
brandNoOPTIONAL but recommended. Pass the brand context inline (name, audience, voice, pillars, competitors, visualCues). Cheapest path: $0, instant. ChiefLab uses this directly instead of scanning. See chiefmo_diagnose_marketing for the full brand schema.
channelsNoWhich launch channels to generate. Default: ["linkedin", "x", "product_hunt", "email", "landing_hero"]. Pass a subset to skip channels you don't want.
tenantIdNoOptional. The end-user this launch is for. Loads stored brand context + voice samples + memory if you've called chieflab_set_tenant_context before.
outputModeNo"context" (default, ~$0): per-channel briefs your agent's LLM renders into final copy. "full": ChiefLab generates the actual posts server-side (premium credits, slower, useful when caller has no LLM).
productUrlNoThe URL of the product being launched. Used as primary signal for positioning + brand extraction. Pass either this or productDescription (or both).
templateIdNoOptional. Apply a saved launch template (chieflab_list_launch_templates → chieflab_save_launch_template). Template channels / launchAngle / brand / repoContext fill missing values; caller-provided values still win.
webhookUrlNoOptional HTTPS URL ChiefLab POSTs to when launch pack is ready, when actions are approved, and when posts go live. Signed with X-ChiefLab-Signature.
launchAngleNoOptional. If you already know the launch angle ("ship-friend launch", "product hunt blitz", "founder-led whisper"), pass it. Otherwise ChiefLab derives one from the product.
productTypeNoOptional override for product-type detection (devtool | saas | ai_agent | marketplace | internal_tool | consumer). Without this, ChiefLab auto-detects from URL + description + repoContext signals.
repoContextNoSTRONGLY RECOMMENDED for agent-built products. Pass the codebase context for the product being launched — the more grounding ChiefLab has, the less generic the launch pack. Without this, outputs read like 'launch any SaaS.' With it, outputs reference the actual product. Coding agents (Cursor / Claude Code / Codex / Bolt / Lovable) should gather this from the user's repo before calling.
scheduleForNoOptional ISO 8601 timestamp. If passed, publish actions schedule for this time instead of immediate-on-approval. UTC.
imagesNeededNoHow many on-brand launch graphics to generate via Gemini. Default: 1 (one hero graphic). 0 to skip image gen entirely. Image gen is metered separately (~$0.04 cost / ~$0.10 retail per image).
approverEmailNoOptional. The human who will approve publish actions. Surfaced in the onboarding checklist.
idempotencyKeyNoOptional dedupe key. Critical for launches — prevents double-publish if the agent retries.
productDescriptionNoOne-paragraph description of what the product is, who it's for, and why it matters. Use when productUrl isn't ready yet (pre-launch, private repo, internal tool).
Behavior5/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description fully covers behavior: it explains the return structure (launchPack, publishActions, agentGuide, reviewUrl), idempotency for deduplication, cost implications for images, webhook callbacks, and the difference between output modes. No behavioral aspects are hidden.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is well-structured with front-loaded purpose, clear sections, and bullet points. Every sentence adds essential context without being verbose. It balances conciseness with completeness, making it easy for an agent to parse quickly.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness5/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's complexity (16 parameters, nested objects, no output schema), the description is complete. It details the return type extensively, covers edge cases like pre-launch scenarios, and provides fallback options. The sibling tool differentiation and usage guidelines ensure the agent has full context.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters5/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

All 16 parameters have descriptions in the schema (100% coverage). The description adds significant value beyond the schema, such as explaining that icp overrides other context, brand is the cheapest path, repoContext is strongly recommended, and productUrl is the primary signal. This extra context greatly aids correct invocation.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly identifies the tool as a 'Repo-aware launch operator for agent-built products,' specifying the action (launch) and resource (product). It distinguishes from the sibling tool chiefmo_diagnose_marketing, which is for diagnosing existing marketing programs, ensuring the agent understands when to use this tool.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines5/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides explicit trigger phrases, context for when to call (user just finished building/shipping/deploying), and contrasts with alternative tools. It includes detailed guidance for coding agents to gather repoContext before calling and describes the IDE-native flow for approval, making usage crystal clear.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_list_actionsCInspect

List ChiefMO proposed actions, optionally filtered by run id.

ParametersJSON Schema
NameRequiredDescriptionDefault
runIdNo
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description must disclose behavioral traits. It only states 'list' and 'optionally filtered by run id,' but omits details like sorting, pagination, permissions, or what constitutes a 'proposed action.' This is insufficient for safe agent selection.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Description is a single sentence with 9 words, efficiently conveying the core function without redundancy. No unnecessary content.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's simplicity (0-1 parameter, no output schema), the description is minimally adequate but lacks behavioral context. It would benefit from clarifying what 'proposed actions' are and any pagination or ordering defaults.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters2/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The input schema has one parameter (runId) with no description, resulting in 0% schema coverage. The description adds only that filtering is optional, which is trivial. It does not explain the format or expected values of runId.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

Description clearly states the verb 'List' and resource 'ChiefMO proposed actions' with an optional filter by run ID. However, it does not distinguish this tool from sibling tools like chiefmo_list_runs or chiefmo_get_action that also deal with actions or runs.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance on when to use this tool versus alternatives such as chiefmo_get_action (for a single action) or chiefmo_list_runs (for runs). The implied usage is for listing actions, but no exclusions or context are provided.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_list_assetsAInspect

List ChiefMO generated assets, optionally filtered by run id.

ParametersJSON Schema
NameRequiredDescriptionDefault
runIdNo
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations provided, but description indicates a read-only listing operation. For a simple list tool, minimal disclosure is adequate.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Single sentence with no waste, front-loaded with the action.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Adequate for a simple list tool with one optional parameter; no output schema needed for basic understanding.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Adds meaning by explaining the runId parameter's purpose (optional filter), compensating for the schema's lack of parameter descriptions.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Clearly states it lists ChiefMO generated assets with optional filtering by run id. Distinguishes from siblings like get_asset and list_runs.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Implies usage through the optional filter but lacks explicit context on when to use this tool vs siblings.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_list_runsBInspect

List recent ChiefMO runs.

ParametersJSON Schema
NameRequiredDescriptionDefault

No parameters

Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations provided. The description implies a read operation but does not state read-only behavior, side effects, rate limits, or return format. The burden falls entirely on the description.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Extremely concise at three words, with no redundant information. Every word matters.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the simplicity (no parameters, no output schema), the description is adequate but lacks context on what constitutes a 'recent' run, expected output fields, or pagination. Could be more complete for a novice agent.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The input schema has zero parameters, so schema coverage is 100%. No additional parameter information needed; description adds nothing beyond what schema provides, but baseline is 4.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the verb 'list' and resource 'runs', with 'recent' adding temporal scope. It distinguishes from sibling 'chiefmo_get_run' but does not differentiate from other list tools like 'chiefmo_gtm_run_list'.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance on when to use this tool versus alternatives (e.g., 'chiefmo_get_run' or 'chiefmo_gtm_run_list'). Usage context is implied but not explicit.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_list_skillsAInspect

List internal ChiefMO marketing skills available to the agent.

ParametersJSON Schema
NameRequiredDescriptionDefault

No parameters

Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Without annotations, the description carries the full burden. It indicates a read-only listing operation but does not disclose potential side effects, auth requirements, or rate limits. The phrase 'available to the agent' adds minor context, but overall transparency is minimal.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, front-loaded sentence of eight words. Every word is necessary and contributes to understanding. There is no wasted content.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness5/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's simplicity (no parameters, no output schema), the description fully covers what the tool does: it lists a specific set of items. The context of 'internal ChiefMO' and 'available to the agent' completes the picture for this listing operation.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters5/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The tool has zero parameters and 100% schema coverage. The description adds value beyond the schema by specifying the scope ('internal ChiefMO marketing skills') and audience ('available to the agent'), which is more informative than an empty schema alone.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description 'List internal ChiefMO marketing skills available to the agent' clearly states the verb (List) and the specific resource (internal ChiefMO marketing skills). It distinguishes this tool from sibling list tools (e.g., list_actions, list_assets) by naming 'skills' as a unique resource.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

No guidance is provided on when to use this tool versus alternatives like list_actions or list_runs. The description lacks explicit when-to-use or when-not-to-use instructions, leaving the agent to infer usage from the tool name alone.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_magic_editAInspect

USE WHEN the user wants a quick tweak to existing marketing copy — 'shorter', 'warmer tone', 'change CTA to question', 'more direct', 'remove the third bullet', 'make it founder-led not corporate'. Apply a one-shot edit to a piece of copy. Default outputMode 'context' returns a small edit-brief your agent's LLM applies (~$0 token cost). Pass outputMode 'full' to have ChiefLab re-render server-side (premium credits). Use for rapid iteration without re-running the full diagnose flow.

ParametersJSON Schema
NameRequiredDescriptionDefault
brandNoOptional inline brand context (same shape as chiefmo_diagnose_marketing.brand). Cheaper than tenantId lookup if your agent already has the data.
originalYesThe text to edit. Whatever your agent's LLM produced last time, paste it here.
tenantIdNoOptional. If passed, ChiefLab applies the tenant's brand voice + voice samples + memory to the edit instruction.
outputModeNo"context" (default, ~$0): returns an edit-brief your LLM applies. "full": ChiefLab re-renders server-side (premium credits, slower).
contentTypeNoOptional. What the original is (linkedin_post, x_post, email, landing_hero, ad_copy, etc.). Helps the brief target platform-specific constraints.
instructionYesWhat to change. Examples: "make it shorter", "warmer tone", "change CTA to a question", "remove the third bullet", "more founder-led, less corporate".
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description carries the full burden. It clearly explains the two output modes ('context' vs 'full'), their costs, and what they return. It also mentions how the tool applies tenant brand voice when tenantId is provided. There is no mention of destructive side effects, but the tool appears safe for editing.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is front-loaded with the key purpose and examples. It is structured with clear sections and uses bullet-like examples. While it is a bit long, every sentence adds value, and it avoids unnecessary fluff.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given that there is no output schema, the description adequately covers return values (edit-brief or re-rendered copy). It explains all six parameters with enough detail for an agent to invoke correctly, including optional ones. The complexity is moderate, and the description is complete.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, so baseline is 3. The description adds value by providing examples for the 'instruction' parameter and explaining the 'outputMode' options. It also clarifies the role of 'brand', 'tenantId', and 'contentType' parameters, enhancing understanding beyond the schema.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description starts with 'USE WHEN the user wants a quick tweak to existing marketing copy' and provides concrete examples like 'shorter', 'warmer tone', making the purpose immediately clear. It distinguishes itself from sibling tools like chiefmo_diagnose_marketing by emphasizing rapid iteration without the full diagnose flow.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description explicitly states when to use the tool ('USE WHEN') and provides examples of instructions. It advises against re-running the full diagnose flow, guiding the agent toward this tool for quick edits. However, it does not explicitly mention when not to use it beyond that.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_measure_launch_resultsAInspect

Pull post-launch metrics + recommend next iteration. USE WHEN it's been ≥24h since chiefmo_publish_approved_post fired and the user asks 'how did the launch perform' or 'what should we do next.' Returns Zernio engagement (likes / comments / views / shares per account) + GA4 traffic if connected + Search Console queries if connected + a structured next-move recommendation brief your LLM renders into a customer-facing summary. The closing tool in the closed loop: build → launch → publish → measure → iterate.

ParametersJSON Schema
NameRequiredDescriptionDefault
runIdYesThe launchId returned by chiefmo_launch_product. Required.
lookbackDaysNoHow many days of history to consider (default 7, max 90).
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description carries full burden. It lists output types (Zernio engagement, GA4 traffic, Search Console queries, recommendation brief) but does not explicitly state whether the tool is read-only or has side effects. The output description is adequate but lacks a clear safety or mutability statement.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single, well-structured paragraph that front-loads purpose, then usage condition, then output, then context. Every sentence adds value; no redundancy. It is concise and easy to parse.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness5/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's complexity (aggregating multiple data sources and generating a recommendation), the description covers purpose, when to use, what is returned, and how it fits into the overall loop. No output schema exists, but the description adequately explains the output structure. The tool is well positioned within its sibling ecosystem.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The description adds context to the parameters beyond the schema: it notes that runId is required and comes from chiefmo_launch_product, and gives default (7) and maximum (90) for lookbackDays. This is valuable usage guidance that the schema alone does not provide.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description starts with a clear verb+object: 'Pull post-launch metrics + recommend next iteration.' It distinguishes itself as 'the closing tool in the closed loop: build → launch → publish → measure → iterate.' The purpose is specific and unambiguous.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description explicitly states when to use: 'USE WHEN it's been ≥24h since chiefmo_publish_approved_post fired and the user asks...' It provides concrete user query triggers. No explicit when-not wording, but the condition implies constraints. A minor gap is not listing alternative tools for other scenarios.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_memory_diffAInspect

Report what the company brain has learned in a window. New repo facts, proof points, channel performance, persona/competitor updates. Use as a weekly review or to show stakeholders why the moat compounds.

ParametersJSON Schema
NameRequiredDescriptionDefault
goalYesWindow description (e.g. 'last 14 days', 'since launch v2').
brandUrlNoOptional brand URL.
sinceIsoNoOptional ISO timestamp; defaults to 30 days ago.
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations exist, so the description carries full burden. It clearly indicates a read-only report operation with no destructive effects, but could mention authentication needs or if any state is affected.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is two sentences: the first specifies the action and content, the second provides usage context. No fluff, optimally concise.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's simplicity (report learned items over a window) and no output schema, the description is largely complete. It could optionally describe the return format, but it's not critical for correct invocation.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100%, and the description does not add meaning beyond the parameter descriptions already provided. Baseline score of 3 is appropriate.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool reports what the company brain has learned over a time window, listing specific content types (repo facts, proof points, etc.). It effectively distinguishes from siblings by focusing on a diff/window report rather than full brain read or extraction.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides use cases like weekly review or stakeholder presentation, giving clear context for when to use. However, it does not explicitly exclude other tools or mention alternatives.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_monitor_competitorsAInspect

Snapshot competitor positioning, recent content, and pricing signals. Returns response-angles per competitor and white-space opportunities for the brand. Re-run monthly; deltas surface in chieflab_competitor_watch.

ParametersJSON Schema
NameRequiredDescriptionDefault
goalYesWhich competitors to track and what to watch for.
brandUrlNoOptional brand URL.
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description must carry the transparency burden. It discloses the tool's output (angles and white-space) and suggests it is a periodic snapshot. It does not contradict any annotations. It could be more explicit about side effects or data handling, but it is adequate.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is two sentences, each with a clear role: first states purpose and output, second gives usage frequency and follow-up. Every sentence earns its place with no waste.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Despite no output schema, the description names the outputs (response-angles, white-space opportunities), which is sufficient for a monitoring tool with simple parameters. It could mention format or frequency, but it is complete enough for an agent to understand.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100%, so the baseline is 3. The description does not add significant meaning beyond the schema: 'goal' is described similarly, and 'brandUrl' is just noted as optional. No additional parameter details are provided.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description uses specific verbs 'Snapshot' and 'Returns' to clearly state the tool's purpose: monitoring competitor positioning, content, and pricing. It distinguishes from siblings by specifying outputs like 'response-angles per competitor' and 'white-space opportunities'.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description advises re-running monthly and mentions where deltas surface, providing clear usage guidance. However, it does not explicitly state when not to use this tool or name alternative tools, leaving some ambiguity.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_plan_content_calendarAInspect

Build a 4-week multi-channel content calendar (channel cadence, deadlines, approval gates, re-plan triggers). Manual mode emits markdown tables; connected mode (Zernio/Resend) auto-fires approved items on their scheduled dates.

ParametersJSON Schema
NameRequiredDescriptionDefault
goalYesCalendar scope — e.g. 'Q2 launch month' or '4 weeks of pre-PMF founder content'.
brandUrlNoOptional brand URL.
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations provided; the description reveals behavioral traits like manual mode output and connected mode auto-firing, but lacks details on side effects, auth needs, or rate limits.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two sentences, no redundant information, front-loaded with the core action. Highly efficient.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Covers output formats and modes adequately. Lacks details on failure handling or setup steps, but sufficient for a planning tool with moderate complexity.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100% with clear descriptions for both parameters. The description adds no extra semantic meaning beyond what the schema already provides, so baseline score of 3.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool builds a 4-week multi-channel content calendar with specific features like deadlines and approval gates. It distinguishes itself from sibling tools by being the only content calendar planner.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage for planning content calendars but does not explicitly state when to use this tool versus alternatives or what prerequisites are needed.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_post_launch_reviewAInspect

USE WHEN ≥24h has passed since chiefmo_publish_approved_post fired and the user asks 'how did the launch perform?', 'what worked?', 'metrics from my launch'. Closes the loop: pulls 24h+ engagement (likes / comments / views / shares / follower delta) for posts in the launch run + recommends next iteration (rewrite, refresh creative, double down, pause). Returns per-account analytics + top posts + best-time-to-post + a structured next-move recommendation brief.

ParametersJSON Schema
NameRequiredDescriptionDefault
runIdYesThe launchId returned by chiefmo_launch_product. Used to scope analytics to posts from that launch.
accountIdsNoOptional. Specific Zernio accountIds to pull analytics for. Defaults to all accounts that posted in this launch.
outputModeNo"context" (default): returns metrics + a brief your LLM renders into the customer-facing recommendation. "full": ChiefLab generates the recommendation server-side.
lookbackDaysNoHow many days of history to consider (default 7, max 90).
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations provided, so description carries full burden. It describes pulling engagement metrics, returning analytics, and recommending next steps. It clearly indicates this is an analytical/review tool, but does not explicitly state if it has side effects or is read-only. The behavior is well-covered.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Three sentences, front-loaded with usage condition. Each sentence adds value: condition, what it returns, output details. Could be slightly more concise but well-structured.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

With 4 params (all described), no output schema, no annotations, the description adequately explains inputs and outputs (engagement metrics, recommendations). It is complete enough for an agent to invoke correctly.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100% (baseline 3). Description adds extra meaning: runId is from chiefmo_launch_product, accountIds defaults to all, outputMode has 'context' vs 'full', lookbackDays default 7 max 90. This provides helpful context beyond the schema.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool's purpose: post-launch review after ≥24h, pulling engagement metrics and recommending next moves. The name 'chiefmo_post_launch_review' is a specific verb+resource combination, and it distinguishes itself from siblings like chiefmo_launch_product and chiefmo_measure_launch_results.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines5/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Explicitly states when to use ('USE WHEN ≥24h has passed since chiefmo_publish_approved_post fired and the user asks...'), providing clear context and implied alternatives (e.g., not for immediate analysis).

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_prepare_next_moveAInspect

USE WHEN the user has just published a launch (or ≥24h post-launch) and asks 'what's next?', 'follow-up post?', 'iterate on this', 'plan day 2'. P13 — turns one launch into a 7-day launch mode: drafts a queue of follow-up actions (day 2 reply post, day 3 case study, day 5 metrics share, day 7 retrospective) grounded in the prior launch's runId, the per-tenant brand voice + repo facts + channel performance from the P9 company brain, and the prior recommendation (so the next move doesn't repeat the last one). Returns a draft queue + signed reviewUrl + the standard agentGuide block. Each next-move action lands in awaiting_approval; nothing fires without explicit human approval.

ParametersJSON Schema
NameRequiredDescriptionDefault
runIdYesThe prior launchId returned by chiefmo_launch_product. The prepared queue grounds in this run's measurement + memory.
horizonNo"day_2" | "day_3" | "week" — how far to plan. Default: "week" (7-day mode).
channelsNoOptional. Restrict next-moves to these channels. Defaults to the channels of the prior launch.
workspaceIdNoOptional workspace id.
Behavior5/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description fully carries the behavioral transparency burden. It discloses that the tool drafts a queue, returns a review URL and agent guide, and crucially states that nothing fires without explicit human approval, which is important safety information.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is fairly concise (about 5 sentences) and structured logically, starting with usage condition, then what it does, then return values. It could be slightly shorter but still efficient.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness5/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's complexity (planning a multi-day queue) and the absence of an output schema, the description thoroughly explains the inputs, behavior, and outputs, making it complete for an AI agent to understand.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, so parameters are already documented. The description adds value by explaining that runId is the prior launchId from chiefmo_launch_product, horizon defaults to 'week', and channels default to prior launch channels, providing context beyond the schema.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description uses a specific verb ('prepare next move') and resource ('launch follow-up actions'), and distinguishes itself from siblings by explicitly mentioning it turns a launch into a 7-day mode, which is unique among the listed tools.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides an explicit 'USE WHEN' condition with example user queries and states it requires a prior launch (runId). It lacks explicit alternatives or when not to use, but the context is clear enough for correct invocation.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_prepare_weekly_reportAInspect

Compose a weekly marketing report with KPI deltas, what changed and why, wins, risks, and recommended next 3 actions. Grounded in connector snapshots when available.

ParametersJSON Schema
NameRequiredDescriptionDefault
goalYesReporting period and focus (e.g. 'last week's marketing across all channels').
brandUrlNoOptional brand URL for context.
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description carries the full burden. It explains the report contents but does not disclose behavioral traits like whether it reads existing data, generates new content, or requires specific permissions. The absence of side-effect or limitation details leaves gaps.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two short sentences, front-loaded with the core purpose. No redundant words; every part adds value. Highly efficient.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a tool with 2 parameters and no output schema, the description covers the high-level purpose and contents. However, it lacks details about output format (e.g., text, JSON) and what happens when connector snapshots are unavailable. Could be more complete for agent decision-making.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100% (both goal and brandUrl described). The description adds minimal value beyond the schema, essentially paraphrasing. The parameter descriptions in schema already provide sufficient semantics.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool composes a weekly marketing report with specific contents (KPI deltas, changes, wins, risks, next actions). It distinguishes itself from analytical siblings like chieflab_measure and chiefmo_synthesize_marketing_analytics by focusing on a structured weekly report format.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies use for weekly reporting but provides no explicit guidance on when to use versus alternatives, nor when not to use. The mention of 'grounded in connector snapshots when available' hints at data source requirements but lacks clarity on prerequisites.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_publish_approved_postAInspect

Publish an approved social post (LinkedIn / X / Threads / Instagram / Facebook / Bluesky / TikTok) via Zernio. USE WHEN the user has approved a publishAction from chiefmo_launch_product on the reviewUrl and you need to fire the post. Strict approval gate: returns { reason: 'requires_approval', reviewUrl } if the action isn't approved yet. Once executed, returns the live platform post URL. Money/external-system action — never call this without an approved actionId.

ParametersJSON Schema
NameRequiredDescriptionDefault
contentYesThe final rendered post text (your LLM rendered this from the brief in the launch pack).
actionIdYesThe publishAction.id from chiefmo_launch_product. Required.
mediaUrlsNoOptional. Image/video URLs (or base64 data URLs from launch graphics).
platformsYesWhich platforms + Zernio accountIds to publish to. Get accountIds from chieflab_list_publish_accounts.
scheduleAtNoOptional ISO 8601. Schedule for later instead of immediate.
Behavior5/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

Discloses the approval gate behavior (returns requires_approval if not approved) and the success result (live platform post URL), and explicitly labels it as a 'money/external-system action' without contradictions.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Concise yet comprehensive; every sentence adds value, with clear structure: purpose, usage trigger, approval gate warning, output summary.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness5/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Fully covers the approval workflow, prerequisites, output format, and safety warnings despite no output schema. Complete for the tool's complexity.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters5/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

All 5 parameters have schema descriptions, and the tool description adds crucial context: actionId source, content rendering, mediaUrls origin, platforms source, and scheduleAt format.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool publishes approved social posts via Zernio across multiple platforms, distinguishing it from sibling tools like chiefmo_send_approved_email and chiefmo_launch_product.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines5/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Explicitly states when to use ('USE WHEN the user has approved a publishAction from chiefmo_launch_product') and when not to use ('never call this without an approved actionId'), providing clear context.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_record_brand_voiceAInspect

Update the brand-voice facet of the company brain: banned phrases (never use), preferred claims (always available), tone descriptors, voice examples. Run after rejecting a draft or after a founder says 'don't write it like that.' Future runs inherit these constraints automatically.

ParametersJSON Schema
NameRequiredDescriptionDefault
tenantIdNo
visualStyleNo
workspaceIdNo
bannedPhrasesNoLowercase phrases future drafts must never contain.
voiceExamplesNo
preferredClaimsNoSpecific claim phrasings this brand prefers.
toneDescriptorsNo
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description carries the full burden. It explains the tool updates the brand-voice facet and that future runs inherit constraints. However, it does not disclose whether changes are additive or replace existing settings, any permission requirements, or other behavioral traits like idempotency or side effects.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is concise, with three short sentences that front-load the core purpose and usage guidance. Every sentence adds value without redundancy or unnecessary detail.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the complexity (7 parameters, nested objects, no output schema, no annotations), the description is incomplete. It fails to mention the visualStyle parameter and does not explain the structure of voiceExamples or toneDescriptors beyond naming them. The tool's overall behavior and return values are not covered.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The description adds some value by listing the key parameters (banned phrases, preferred claims, tone descriptors, voice examples) with brief purpose labels. However, it does not cover all parameters (e.g., visualStyle is omitted), and the schema already has descriptions for two parameters. Given low schema coverage (29%), the description partially compensates but is insufficient for a full understanding.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states it updates the brand-voice facet of the company brain, listing the specific components (banned phrases, preferred claims, tone descriptors, voice examples). It uses a specific verb and resource, and the context differentiates it from sibling tools that deal with other facets.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description explicitly provides usage context: 'Run after rejecting a draft or after a founder says 'don't write it like that.' It also states that future runs inherit constraints, giving clear guidance on when to use this tool. However, it does not mention when not to use it or name alternative tools.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_record_channel_resultBInspect

Record one channel-performance row (channel × hook × format × CTA × outcome × metric). Powers the 'what's worked / hasn't' block in future plans. Outcome is one of: won | lost | inconclusive | killed.

ParametersJSON Schema
NameRequiredDescriptionDefault
ctaNo
hookNo
notesNo
formatNopost | thread | carousel | video | image | long_form | newsletter
channelYeslinkedin | x | email | blog | ad | reddit | threads | instagram | product_hunt
outcomeNowon | lost | inconclusive | killed (default: inconclusive)
baselineNo
tenantIdNo
sourceRunIdNo
workspaceIdNo
primaryValueNo
primaryMetricNo
sourceAssetIdNo
sourceActionIdNo
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description must disclose behavioral traits. It only mentions the outcome enum and that it records a row. It does not cover whether the operation overwrites data, required permissions, side effects, or default values beyond the schema's outcome default. The description carries the full burden but offers minimal behavioral context.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is two sentences, each earning its place. The first sentence is dense and informative, listing the core dimensions. The second sentence clarifies the outcome enum. No wasted words.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given 14 parameters, no output schema, and no annotations, the description is too brief. It doesn't explain the relationship between parameters, defaults beyond outcome, or what happens after a row is recorded. For a data entry tool, the description should provide more operational context.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters2/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is low (21%), so the tool description should compensate. It adds value by naming key parameters (channel, hook, format, CTA, outcome, metric) but does not explain the remaining 8 parameters (e.g., baseline, primaryMetric, sourceAssetId). The description partially clarifies the content but leaves many parameters ambiguous.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the action (record) and the resource (channel-performance row), specifying the key dimensions (channel, hook, format, CTA, outcome, metric) and its purpose (powers future plans). It distinguishes itself from sibling tools by being a dedicated record operation for individual performance data.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description conveys when to use the tool (recording a performance row for future analysis) but does not explicitly address when not to use it or provide alternatives among related sibling tools like chiefmo_measure_launch_results or chiefmo_list_runs. The guidance is adequate but lacks exclusion criteria.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_record_personaAInspect

Capture an audience/persona memory entry: ICP, segments, top objections, proof points, banned claims, winning angles. Future runs ground against this memory. Use early in onboarding, then refresh when your understanding shifts.

ParametersJSON Schema
NameRequiredDescriptionDefault
goalYesPersona context — what you've learned about who buys / signs up.
brandUrlNoOptional brand URL.
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description carries the full burden. It indicates that future runs 'ground against this memory,' implying stored data affects behavior, but does not disclose details about permissions, idempotency, or data limits.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two sentences, front-loaded with the tool's purpose and key fields, followed by usage advice. Every sentence earns its place with no filler.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's simplicity (2 params, no output schema), the description covers purpose, usage timing, and data scope. It is slightly ambiguous about whether calling again updates or creates a new entry, but overall sufficient.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100% with descriptions for both parameters, so baseline is 3. The description adds value by elaborating on the 'goal' parameter with examples (ICP, segments, etc.), going beyond the schema's brief label.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool captures an audience/persona memory entry and lists specific fields (ICP, segments, top objections, etc.), which distinguishes it from sibling tools like chiefmo_record_brand_voice or chieflab_remember_preferred_channel.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description explicitly advises 'Use early in onboarding, then refresh when your understanding shifts,' providing clear timing guidance. However, it does not compare this tool to alternative memory tools or state when not to use it.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_record_proofAInspect

Capture a structured proof asset (testimonial, metric, customer quote, case study, launch result, review, press) into the company brain. Future runs can cite it instead of inventing claims.

ParametersJSON Schema
NameRequiredDescriptionDefault
bodyYesVerbatim quote or 1-2 sentence summary.
kindYestestimonial | metric | customer_quote | case_study | launch_result | review | press
titleNo
tenantIdNo
sourceUrlNo
metricNameNoRequired when kind=metric or launch_result.
metricUnitNousers | USD | % | pct_pts | etc.
observedAtNoISO date; defaults to now.
attributionNoWho/where (e.g. 'Jane Doe, CTO at Acme'; 'Reddit /r/SaaS').
metricValueNo
workspaceIdNo
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description must cover behavioral traits. It indicates a write operation and lists proof kinds, but omits details on mutation (e.g., overwriting), authentication needs, or side effects like duplicate handling.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is two sentences, front-loads the verb and resource, and contains no superfluous words, making it efficient and easy to parse.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a tool with 11 parameters and no output schema or annotations, the description covers the core function and future utility adequately, though it lacks detail on optional parameters and response format.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 55%, and the description adds minimal parameter-specific meaning beyond listing kinds in the schema. It does not explain optional parameters like tenantId or workspaceId, so it meets baseline but does not compensate for gaps.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the action ('Capture a structured proof asset') and the resource ('into the company brain'), distinguishing it from sibling tools like chiefmo_record_brand_voice by specifying the types of proofs (testimonial, metric, etc.).

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description explains the purpose ('Future runs can cite it instead of inventing claims'), implying when to use it, but does not explicitly state when not to use or provide direct comparisons to alternative tools.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_record_repo_factBInspect

Persist one structured product fact (feature / route / pricing / integration / changelog / screenshot / metric / constraint) extracted from the repo. Older facts with the same (kind, name) are auto-superseded. Run on every launch so the brain stays current with the codebase.

ParametersJSON Schema
NameRequiredDescriptionDefault
kindYesfeature | route | pricing | integration | changelog | screenshot | metric | constraint
nameYes
valueNo
detailsNo
tenantIdNo
sourceFileNo
workspaceIdNo
sourceCommitNo
sourceRepoUrlNo
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

The description discloses the key behavior of auto-superseding same (kind, name) facts, which is helpful. However, with no annotations, it fails to disclose other traits like failure modes, permissions, or idempotency.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is two sentences, front-loaded with purpose and key behavior, with no wasted words.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given 9 parameters, no annotations, no output schema, and the complexity of persisting structured facts, the description is too minimal. It omits details on optional parameters, expected response, and how to use nested objects.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters2/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is only 11%, yet the description only explains the (kind, name) identifier role. It does not clarify the meaning of other parameters like value, details, or sourceFile, leaving significant ambiguity.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states it persists a structured product fact and lists the kinds. However, it does not differentiate from sibling tools like chiefmo_extract_repo_facts or chieflab_brain_read, which could cause confusion.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides context that older facts are auto-superseded and should be run on every launch, but does not explicitly state when to use this tool instead of alternatives or when not to use it.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_reject_runBInspect

Reject a ChiefMO run that is awaiting approval.

ParametersJSON Schema
NameRequiredDescriptionDefault
idYes
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description must carry the full behavioral disclosure. It only states the basic action without mentioning side effects, reversibility, or authorization requirements. This is insufficient for a mutation tool.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is a single short sentence with no redundancy. It is front-loaded and efficient, though it sacrifices detail.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the low complexity (1 parameter, no output schema, no annotations), the description covers the essential purpose but omits preconditions (e.g., run must be in 'awaiting approval' status) and return value. It is minimally adequate.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters2/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The single required parameter 'id' is not explained in the description; the schema only shows it is a string. With 0% schema description coverage, the description should add meaning like 'the ID of the run to reject,' but it does not.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the action ('Reject'), the object ('a ChiefMO run'), and the condition ('that is awaiting approval'). It effectively differentiates from the sibling tool 'chiefmo_approve_run'.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage when a run is pending approval, but does not explicitly state when to use it over alternatives or provide any prerequisites. The context from sibling tools indicates it is the reject counterpart, but no direct guidance is given.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_render_quote_cardAInspect

Render a deterministic on-brand quote card as PNG using Era (HTML/CSS layout + Playwright render), not diffusion. USE WHEN the user asks for a quote card, founder quote graphic, testimonial graphic, text-heavy launch visual, or brand-precise social graphic. Takes brand context and exact quote text, returns a PNG dataUrl plus layout metadata. Local pipeline; no publish/send side effect.

ParametersJSON Schema
NameRequiredDescriptionDefault
brandYesBrand context. Supports { name, voice, palette: { background, foreground, accent }, typography: { heading, body }, markUrl } or legacy visualCues.palette/type.
quoteYesExact quote text to render. Era renders this verbatim.
returnHtmlNoWhether to include the generated HTML in the response. Default false to keep MCP payloads smaller.
attributionYesWho said it, or a short source label.
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full burden. It discloses that the tool is deterministic, uses a specific tech stack (Era/Playwright), has no side effects (local pipeline, no publish/send), and returns a PNG dataUrl plus metadata. This is transparent, though it could mention any prerequisites or performance characteristics.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Three sentences with clear front-loading: first sentence states core action and technology, second lists use cases, third summarizes inputs/outputs and side effects. Every sentence adds value; no wasted words.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a tool with 4 parameters (3 required, nested objects) and no output schema, the description covers the key aspects: what it does, inputs, output type (PNG dataUrl + metadata), and side-effect status. It could be slightly more detailed about the output format but is sufficient.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, so baseline is 3. The description adds minor context (e.g., 'Era renders this verbatim' for quote, and explains the brand parameter accepts legacy fields). However, it doesn't significantly enhance understanding beyond the schema descriptions.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool renders a deterministic quote card as PNG using Era (HTML/CSS + Playwright), not diffusion, and lists specific use cases (quote card, founder quote graphic, etc.). This distinguishes it from sibling tools like chiefmo_generate_ad_variants or chiefmo_create_social_graphics.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description explicitly states when to use the tool: for quote cards, testimonial graphics, text-heavy visuals, or brand-precise social graphics. It lacks explicit 'when not to use' or alternatives, but the positive use cases are clear and specific.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_repurpose_contentAInspect

Take a single source asset (launch post, blog, demo, customer proof) and emit platform-tuned variants for LinkedIn, X thread, Threads, and email. Each variant leads with a different hook — not paraphrases. Manual mode pastes; connected mode (Zernio/Resend) can publish on a cadence after approval.

ParametersJSON Schema
NameRequiredDescriptionDefault
goalYesDescribe the source content and what you want repurposed.
brandUrlNoOptional brand URL.
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations provided; description carries full burden. It explains output behavior (different hooks, platform tuning) and two modes (manual paste, connected publish after approval). However, lacks details on side effects, storage, or permissions.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two sentences, no fluff, front-loaded with purpose. Could be more structured but efficiently conveys key information.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

No output schema and description fails to state what the tool returns (e.g., variant texts). For a tool with moderate complexity, this is a significant gap.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100% with descriptions for both parameters. The description adds context: 'source content and what you want repurposed' for goal and 'Optional brand URL' for brandUrl, exceeding schema alone.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

Clearly states the tool takes a single source asset and produces platform-tuned variants for LinkedIn, X, Threads, and email, each with a different hook, distinguishing it from sibling tools.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Describes manual vs connected modes but does not explicitly state when to use this tool over similar siblings like chiefmo_generate_social_posts or chiefmo_draft_blog_post. Missing when-not or alternative guidance.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_send_approved_emailAInspect

Send an approved launch email via Resend. USE WHEN the user has approved an email publishAction from chiefmo_launch_product and you need to fire the send. Strict approval gate (same shape as chiefmo_publish_approved_post). from MUST be on a verified Resend domain — check chieflab_list_email_senders first. Single recipient or short list (≤50). Money/external-system action — once sent cannot be unsent.

ParametersJSON Schema
NameRequiredDescriptionDefault
toYesRecipient(s). Single email or array (≤50).
fromYesSender. Must be a verified Resend domain address (e.g. 'Brand <hi@yourdomain.com>').
htmlNoHTML body (preferred for inboxing).
textNoPlain-text fallback (recommended).
replyToNoOptional reply-to address(es).
subjectYes
actionIdNoThe publishAction.id from chiefmo_launch_product. Required for the approval gate to fire.
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description carries the full burden. It explicitly calls out the action as a 'Money/external-system action' and states 'once sent cannot be unsent', indicating destructive side effects. It also mentions the approval gate requirement. However, it does not detail rate limits, authentication specifics, or error conditions beyond domain verification.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is four sentences with a clear front-loaded purpose statement. Every sentence adds value without repetition. The structure flows logically: what it does, when to use, prerequisites, and a warning. No wasted words.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given 7 parameters, no output schema, and no annotations, the description covers the workflow (approval gate, send), preconditions (verified domain, approval), and constraints (≤50 recipients, irreversibility). However, it does not describe the return value (e.g., status, email ID) or error handling steps, which would improve completeness for an agent.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters4/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 86% (6/7 params described). The description adds meaningful context: it emphasizes that 'from' must be on a verified Resend domain and instructs to check senders first, which is not fully captured in the schema's description. It also reinforces the role of 'actionId' for the approval gate. This augments the schema adequately.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool sends an approved launch email via Resend, specifically tied to an email publishAction from chiefmo_launch_product. It explicitly distinguishes its purpose from siblings like chiefmo_publish_approved_post by noting the same approval gate shape but different resource type.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines5/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides explicit when-to-use: after user approves an email publishAction from chiefmo_launch_product. It includes prerequisites (checking verified Resend domain via chieflab_list_email_senders), constraints (≤50 recipients), and a warning about irreversibility. This leaves no ambiguity about the proper invocation context.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_seo_briefAInspect

Generate a Search Console-grounded SEO refresh brief: which pages to refresh, which queries to target, content clusters to build, internal-link opportunities. Manual mode emits a structural brief; with Search Console connected, every recommendation cites actual impression / CTR / position numbers.

ParametersJSON Schema
NameRequiredDescriptionDefault
goalYesSEO question or target (e.g. 'recover ranking on /pricing', 'plan next quarter's content clusters').
brandUrlNoOptional brand URL.
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations, the description carries the burden and discloses mode-dependent behavior (structural brief vs. data-cited brief). It does not mention side effects, authentication, or rate limits.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

Two sentences, no fluff, front-loaded with purpose and key outputs, efficient and informative.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

The description covers purpose, modes, and output nature but does not specify return format or pagination. Given no output schema, a brief mention of output type would improve completeness.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema coverage is 100% and both parameters are described in the schema. The description adds no additional detail beyond what the schema provides for parameter semantics.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool generates an SEO refresh brief grounded in Search Console data, listing specific outputs (pages, queries, clusters, links) and two modes, distinguishing it from sibling tools.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implicitly differentiates modes but does not explicitly state when to use this tool versus alternatives like chiefmo_draft_blog_post or chiefmo_plan_content_calendar.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

chiefmo_synthesize_marketing_analyticsAInspect

Cross-channel marketing analytics synthesis. Pulls social / email / blog / GA4 / Search Console signals into a single 'what worked, what didn't, what to do next' story. Ends with 3 prioritized actions and memory updates to feed forward.

ParametersJSON Schema
NameRequiredDescriptionDefault
goalYesReporting period or focus (e.g. 'last 30 days across all channels').
brandUrlNoOptional brand URL.
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description adequately discloses that the tool pulls data from multiple channels, synthesizes it into a story, produces 3 prioritized actions, and updates memory. This covers the key behavioral traits (data retrieval, synthesis, memory writes) but does not detail potential side effects like memory overwriting.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is three sentences with no fluff. It front-loads the core purpose ('Cross-channel marketing analytics synthesis') and efficiently conveys the tool's inputs, process, and outputs.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's moderate complexity (2 params, no output schema, no annotations), the description covers the essential aspects: what it does, what data it uses, what it produces, and the memory update side effect. It lacks specifics on permissions or data freshness but is otherwise complete for an agent to decide when to invoke it.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The input schema already provides 100% coverage with descriptions for both parameters (goal and brandUrl). The description adds no additional semantics beyond the schema, so the baseline score of 3 is appropriate.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description identifies the tool as a cross-channel marketing analytics synthesizer, specifying the channels (social, email, blog, GA4, Search Console) and the output format ('what worked, what didn't, what to do next' story with 3 prioritized actions and memory updates). This clearly distinguishes it from sibling tools like chiefmo_diagnose_marketing, which likely focus on troubleshooting rather than synthesis.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines3/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description implies usage for periodic reporting (e.g., 'last 30 days') but does not explicitly state when to use this tool versus alternatives like chiefmo_diagnose_marketing or chiefmo_measure_launch_results. No exclusions or when-not-to-use guidance is provided.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

Discussions

No comments yet. Be the first to start the discussion!

Try in Browser

Your Connectors

Sign in to create a connector for this server.

Resources