Skip to main content
Glama
127,390 tools. Last updated 2026-05-05 15:21

"An automation tool for workflows and integrations (n8n software)" matching MCP tools:

  • DEPRECATED — use create_tmb_job instead. Posts a job as an on-chain TMB contract with platform resolver and dispute protection. This tool returns an error directing you to create_tmb_job.
    Connector
  • Retrieve an AWS agent skill — domain-specific expertise that transforms you into a specialist for a particular AWS domain. Skills provide workflows, context, best practices, decision frameworks and step-by-step procedures. A skill may include reference files (architecture docs, schemas, examples) and deterministic workflows for sub-tasks that require exact execution. ## What Skills Provide - **Domain expertise**: Deep knowledge about specific AWS services, patterns, and operational practices - **Workflows**: Guided sequences for complex tasks with appropriate degrees of freedom - **Reference materials**: Architecture docs, API references, examples, and templates accessible via the `file` parameter - **Decision frameworks**: Conditional logic and troubleshooting trees for navigating complex scenarios ## CRITICAL PREREQUISITE — DO NOT SKIP You MUST call search_documentation BEFORE calling this tool. NEVER call this tool first. You do NOT know skill names — they are unpredictable identifiers that can only be discovered through search_documentation results. Guessing or fabricating a skill_name WILL fail. ## REQUIRED WORKFLOW (no exceptions) 1. FIRST: Call search_documentation with the user's requirements 2. THEN: Find the result entry that has a skill_name field 3. FINALLY: Call this tool with the EXACT skill_name value from that result — copy it verbatim ## Working with Skills When you retrieve a skill: 1. Read the SKILL.md overview to understand the domain and scope 2. Follow the workflows and guidance in the skill body 3. When the skill references additional files (e.g., `[architecture](references/architecture.md)`), retrieve them using this same tool with the `file` parameter 4. Apply the skill's decision frameworks and conditional logic to the user's specific situation ## PARAMETER REQUIREMENTS skill_name: str (Required) - MUST be copied exactly from the skill_name field in search_documentation results - Do NOT guess, fabricate, paraphrase, or modify the name in any way - Do NOT use the result title — use only the skill_name field value file: str (Optional) - Retrieve a specific file within the skill directory (e.g., "references/architecture.md") - Use this when the SKILL.md body links to reference files - If omitted, returns the main SKILL.md file ## IF SKILL NOT FOUND If you get an error, you likely guessed the name. Call search_documentation first to discover it. The error response will include a list of available files for the skill. ## Returns The skill content — either the main SKILL.md with domain expertise, workflows, and guidance, or a specific reference file when the `file` parameter is provided.
    Connector
  • Lists every automation configured on a perspective with its trigger, channel (sensitive details redacted), execution mode, enabled state, schedule description, and recent error/success metadata. Behavior: - Read-only. - Errors when the perspective is not found or you do not have access. - Sensitive parts of channel delivery (e.g., webhook auth headers, full URLs) are redacted before being returned. - has_error / last_error / last_error_at / failure_count appear only when there have been recent failures. When to use this tool: - Auditing what's wired up on a perspective before adding more automations. - Finding an automation_id to feed into automation_update, automation_delete, or automation_test. - Diagnosing a failing automation via last_error / failure_count. When NOT to use this tool: - Creating a new automation — use automation_create. - Toggling enabled or changing config — use automation_update. - Verifying delivery actually works — use automation_test.
    Connector
  • Updates fields on an existing automation. Pass a partial updates object with only the fields you want to change; omitted fields are preserved. Toggling enabled or changing schedule/channel/condition takes effect on the next scheduled run. Behavior: - Saves the change to the same automation record. Scheduled automations with an active workflow are restarted on update so the next run picks up the latest config. - Errors when the perspective or automation is not found, or you do not have access. - Webhook URLs in updates are validated. For HubSpot, the workspace's HubSpot connection is re-checked — errors with "Could not resolve HubSpot portal ID — please reconnect HubSpot" if disconnected. - For scheduled automations: changes to channel, condition, execution mode, instruction, or message template apply starting from the next run, not the one currently in flight. When to use this tool: - Toggling enabled on or off (also pauses/resumes scheduled sends). - Changing schedule, channel, condition, instruction, or message_template on a live automation. When NOT to use this tool: - Removing the automation entirely — use automation_delete. - Verifying a config change actually delivers — follow up with automation_test. - Listing what's configured — use automation_list.
    Connector
  • [READ] Search the Layer 3 curated directory of MCP servers and agent-work tools. The directory has 30 entries across three vetting tiers — `first-party` (operated by the swarm.tips DAO), `vetted` (third-party, we've used + verified), `discovered` (cataloged from public sources, not yet exercised). Filter by `query` (substring vs name/description/tags), `category` (substring), and `tier`. Results sort first-party → vetted → discovered. The same directory powers swarm.tips/discover; this tool exposes it programmatically. Use this when an agent needs to find an MCP server for a capability (DeFi, search, browser automation, etc.) instead of an opportunity (which `discover_opportunities` covers).
    Connector
  • Runs a single end-to-end execution of an existing automation against a mock conversation, returning success/failure plus the channel target and duration. Mirrors a real production firing. Behavior: - Sends REAL messages by default: posts the configured webhook, sends the configured email, posts the Slack message, or writes the HubSpot record. Use override_email (email channels) or override_webhook (webhook channels) to redirect delivery to a safe test target. - Each call fires another real delivery. - Errors when the perspective or automation is not found, or you do not have access. Webhook URLs (configured or override) are validated. - Mock conversation defaults: trust score 85, status complete, "Test Participant" / test@example.com. Override participant_name, summary, and tags via test_data. - Returns success: true also when the automation's condition skips delivery (e.g., tag/trust filter doesn't match the mock). The error field is populated only on real delivery failures. When to use this tool: - Verifying a freshly-created automation actually delivers before relying on it. PREFER override_email/override_webhook to avoid spamming real recipients. - Reproducing a delivery failure surfaced in automation_list (last_error). When NOT to use this tool: - Listing what's configured — use automation_list. - Changing config — use automation_update. - Removing the automation — use automation_delete.
    Connector

Matching MCP Servers

Matching MCP Connectors

  • Creates an automation on a perspective. Triggers: per_interview (fires on every completed conversation) or scheduled (daily/weekly digest). Channels: webhook, email, slack, hubspot. Execution modes: direct (fast, deterministic) or agent (LLM-powered). Behavior: - Each call creates a new automation — even if name/config matches an existing one. - Once enabled, the automation starts firing on real events: per_interview sends on every completed conversation going forward; scheduled sends a real message on the configured cadence (daily/weekly). - Webhook URLs are validated. For HubSpot, the workspace's HubSpot connection is required — errors with "Could not resolve HubSpot portal ID — please reconnect HubSpot" if not connected. - Errors when the perspective is not found or you do not have access. When to use this tool: - The user wants ongoing notifications on every completed conversation (per_interview). - Building a daily/weekly digest delivered to Slack, email, HubSpot, or a webhook (scheduled). When NOT to use this tool: - Trying a one-off send before going live — create the automation, then use automation_test (use override_email / override_webhook to avoid hitting real recipients). - Editing or toggling an existing automation — use automation_update. - Connecting Slack or HubSpot — use integration_manage first; the provider must be connected before slack/hubspot channels work. Example — per-conversation Slack notify: ``` { "perspective_id": "...", "automation": { "name": "Notify Slack", "trigger": { "type": "per_interview" }, "execution_mode": "agent", "channel": { "type": "composio", "delivery_config": { "provider": "slackbot", "tool_slug": "SLACKBOT_SEND_MESSAGE", "params": { "channel": "#research" }, "resource_id": "...", "resource_name": "..." } } } } ``` Typical flow: 1. integration_manage (operation: "list"/"connect") → ensure Slack / HubSpot is connected (only needed for those channels) 2. automation_create → create the automation 3. automation_test (with overrides) → verify delivery before relying on it
    Connector
  • Permanently deletes an automation. Pauses any scheduled sends first, then removes the automation. Behavior: - DESTRUCTIVE and irreversible — the automation cannot be recovered. No undo. Confirm with the user before calling. - Errors when the perspective or automation is not found, or you do not have access. Deleting an already-deleted automation errors as well. - If pausing the scheduled sender fails, the deletion is aborted and you'll get success: false with "Failed to stop running workflow. Please try again." — the automation stays intact in that case. When to use this tool: - The user explicitly asked to remove an automation and confirmed. - Cleaning up a misconfigured automation that automation_test repeatedly fails on. When NOT to use this tool: - The user just wants to pause it temporarily — use automation_update with { enabled: false } instead. - You're not sure which automation_id is correct — confirm via automation_list first.
    Connector
  • Initiates the deletion of a Cloud Composer environment. This is a destructive action that permanently deletes the environment and cannot be undone. Users should be asked for confirmation before proceeding. This tool triggers the environment deletion process, which is a long-running operation that typically takes 10-20 minutes. The tool returns an operation object. Use the `get_operation` tool with the operation name returned by this tool to poll for deletion status.
    Connector
  • Execute an integration action — e.g., send an email via Resend, create a payment via Mollie. The system resolves vault credentials server-side so you never handle API keys directly. The integration must be configured first via setup_integration (not needed for built-in integrations). Call get_integration_schema first to get the exact endpoint name and required input fields.
    Connector
  • Internal NotFair tool-feedback channel. Privately report MCP/tool friction that got in the way of helping the user — unclear descriptions, missing capabilities, clunky workflows, confusing errors, or duplicate tools. This is not customer support, not user feedback, and not a success/quality rating. It is an internal NotFair engineering signal. When tool design gets in the way of a real user task, file one concrete report here so we can fix the tool surface. The user benefits in their next session; every other agent serving every other user benefits too. AUTO-SURFACE THIS WHEN: - A tool description was unclear and you weren't sure how to use it. - You wanted to accomplish something for the user but no tool existed for it. - A workflow took many tool calls when one bulk operation could have replaced them. - An error message returned by a tool didn't help you debug or recover. - Two tools have overlapping purposes and the choice was confusing. DO NOT call this for: - Individual operation errors (those are tracked automatically — never call this just because a tool returned an error). - Confirming that a task succeeded. - Rating your own output quality. - Anything the user explicitly asked you to escalate (use the in-app feedback form for that). Be specific. Reference tools by name and propose a concrete change. Keep yourself to at most 2 calls per session. Submissions go directly to the NotFair team; the user does not see this channel.
    Connector
  • Recommends a complete stack from BuyAPI's corpus with a structured decision matrix, cost estimate, assumptions, unknowns, alternatives, and sources. Use this when the user is starting a project or asks for a complete stack choice. Do not use this for local coding/debugging/docs questions that do not involve software or vendor selection. Do not call vendors.resolve first; this tool handles retrieval and ranking.
    Connector
  • Returns structured information about what the Recursive platform includes: features, AI model details, supported integrations, and what's included at every tier. Use for systematic feature comparison.
    Connector
  • Get a compact intelligence digest for a set of brands — perfect for watchlist summaries, competitive briefings, and daily reports. Returns for each brand: current signal, AI visibility score+trend+grade, key relationship edges (integrations, powered-by, acquisitions), and capabilities. Excludes competitive edges to keep output focused. Args: slugs: List of brand slugs (up to 25). Returns: Dict with "digest" array (one entry per brand) and "missing_slugs".
    Connector
  • Get Arcadia workflow guides and reference documentation. Call this before multi-step workflows (opening LP positions, enabling automation, closing positions) or when you need contract addresses, asset manager addresses, or strategy parameters. Topics: overview (addresses + tool catalog), automation (rebalancer/compounder setup), strategies (step-by-step templates), selection (how to evaluate and parameterize strategies).
    Connector
  • Get the builder workflows — step-by-step state machines for building skills and solutions. Use this to guide users through the entire build process conversationally. Returns phases, what to ask, what to build, exit criteria, and tips for each stage.
    Connector
  • [FREE] See all available free SEO tools and recommended workflows. Call this first. FR: Découvrez les outils SEO gratuits et les workflows recommandés.
    Connector
  • Returns this server's runtime configuration: upload endpoint URL, output file TTL, file size limits, and base64 encoding rules. Call this before working with large files (≥ 4 MB) or when building multi-step workflows that chain tool outputs.
    Connector
  • [READ] Aggregated list of paid services swarm.tips agents can spend on. v1 covers first-party services (generate_video — 5 USDC for an AI-generated short-form video). External spend sources (Chutes inference at llm.chutes.ai/v1, x402-paywalled APIs, etc.) are deferred to follow-up integrations. Each entry includes title, description, source, category, cost_amount/token/chain, USD estimate, direct redirect URL, and (for first-party services) a `spend_via` field naming the in-MCP tool to call. Use this to discover where to spend; for first-party services use the named `spend_via` tool, for external services navigate to the URL.
    Connector