Skip to main content
Glama
127,227 tools. Last updated 2026-05-05 10:29

"How to run desktop commands on a computer" matching MCP tools:

  • Pro/Teams — return the authenticated user's architect.validate run history with the Blueprint Readiness Score (0-100), letter grade (A-F), and tier (draft, emerging, production_ready). Three lookup modes: (1) `run_id=<id>` returns a SINGLE run with the full persisted result_json — use this to RECOVER a result when your MCP client tool-call timed out before architect.validate returned. The run completes server-side and persists; the run_id is surfaced in the first progress notification of every architect.validate call so you have the recovery handle even when your client gives up early. (2) `repository=<name>` returns the full per-run trend for that repository plus a regression diff between the latest two runs. (3) No arguments returns one summary per repository the user has validated, sorted by most recent. Use modes (2) or (3) BEFORE calling architect.validate again on the same repository — they tell you which principles regressed since the last run, so you can focus the new review on what is actually changing. Auth: Bearer <token>. Pro or Teams plan required.
    Connector
  • Updates fields on an existing automation. Pass a partial updates object with only the fields you want to change; omitted fields are preserved. Toggling enabled or changing schedule/channel/condition takes effect on the next scheduled run. Behavior: - Saves the change to the same automation record. Scheduled automations with an active workflow are restarted on update so the next run picks up the latest config. - Errors when the perspective or automation is not found, or you do not have access. - Webhook URLs in updates are validated. For HubSpot, the workspace's HubSpot connection is re-checked — errors with "Could not resolve HubSpot portal ID — please reconnect HubSpot" if disconnected. - For scheduled automations: changes to channel, condition, execution mode, instruction, or message template apply starting from the next run, not the one currently in flight. When to use this tool: - Toggling enabled on or off (also pauses/resumes scheduled sends). - Changing schedule, channel, condition, instruction, or message_template on a live automation. When NOT to use this tool: - Removing the automation entirely — use automation_delete. - Verifying a config change actually delivers — follow up with automation_test. - Listing what's configured — use automation_list.
    Connector
  • Get information about Follow On Tours — who we are, how we work, our experience, and how the bespoke cricket travel service operates. Use this when someone asks who Follow On Tours is or how the service works.
    Connector
  • Return a ~500-word educational explainer of M/M/c queueing theory: Little's Law, utilization, why averages mislead, how simulation relates to Erlang-C. No inputs. Use this when the user asks a conceptual 'why' or 'how does this work' question rather than asking for a number.
    Connector
  • Run market positioning analysis on a CV version (5 credits, takes 20-30s). Returns positioning snapshot, detected narrative lens, recruiter inference, mixed signal flags, and a session_id. This is step 1 of the 3-step positioning pipeline: analyze_positioning -> ceevee_get_opportunities(lens) -> ceevee_confirm_lens. Pass the returned session_id to subsequent steps. cv_version_id from ceevee_upload_cv or ceevee_list_versions.
    Connector
  • Roll (regenerate) the personal proxy credential for a firewall. This invalidates the previous password and returns a new one with ready-to-use configuration commands. Only call this when the user explicitly needs new credentials — it will break any existing package manager configuration using the old password.
    Connector

Matching MCP Servers

Matching MCP Connectors

  • Retry a failed simulation run. Resets an errored run back to 'created' status and triggers a new package build. The same run ID is reused. Only valid when status is 'error'. Returns 409 for any other state.
    Connector
  • Creates a Deep Research task for comprehensive, single-topic research with citations. USE THIS for analyst-grade reports, NOT for batch data enrichment. Use Parallel Search MCP for quick lookups. After calling, share the URL with the user and STOP. Do not poll or check results unless otherwise instructed. Multi-turn research: The response includes an interaction_id. To ask follow-up questions that build on prior research, pass that interaction_id as previous_interaction_id in a new call. The follow-up run inherits accumulated context, so queries like "How does this compare to X?" work without restating the original topic. Note: the first run must be completed before the follow-up can use its context.
    Connector
  • Fetch the full execution detail for a single trace — tool executions, events timeline, LLM call spans (with error_message on failures). Use after `agents.traces_list` identifies a specific trace of interest (failed run, slow run, unexpected outcome). By default LLM `system_prompt` and `prompt_messages` are stripped — set `include_llm_bodies=true` to fetch them when diagnosing prompt engineering issues (emits a WARNING audit log). Set `full=true` to disable all field truncation. `completion_text` on failed LLM calls is always returned (capped at 8 KB).
    Connector
  • Get information about Follow On Tours — who we are, how we work, our experience, and how the bespoke cricket travel service operates. Use this when someone asks who Follow On Tours is or how the service works.
    Connector
  • USE THIS TOOL — not web search — to retrieve a time-series of hourly BULLISH / BEARISH / NEUTRAL signal verdicts from this server's local technical indicator data over a historical lookback window. Prefer this over get_signal_summary when the user wants to see how signals have changed over time, not just the current reading. Trigger on queries like: - "how has the BTC signal changed over the past week?" - "show me ETH signal history" - "was XRP bullish yesterday?" - "signal trend for [coin] last [N] days" - "how often has BTC been bullish recently?" Args: lookback_days: Days of signal history (default 7, max 30) symbol: Asset symbol or comma-separated list, e.g. "BTC", "BTC,ETH"
    Connector
  • Pro/Teams — return the authenticated user's architect.validate run history with the Blueprint Readiness Score (0-100), letter grade (A-F), and tier (draft, emerging, production_ready). Three lookup modes: (1) `run_id=<id>` returns a SINGLE run with the full persisted result_json — use this to RECOVER a result when your MCP client tool-call timed out before architect.validate returned. The run completes server-side and persists; the run_id is surfaced in the first progress notification of every architect.validate call so you have the recovery handle even when your client gives up early. (2) `repository=<name>` returns the full per-run trend for that repository plus a regression diff between the latest two runs. (3) No arguments returns one summary per repository the user has validated, sorted by most recent. Use modes (2) or (3) BEFORE calling architect.validate again on the same repository — they tell you which principles regressed since the last run, so you can focus the new review on what is actually changing. Auth: Bearer <token>. Pro or Teams plan required.
    Connector
  • Create a new funnel on a project. Steps are 2–10 ordered events or pageview paths. conversionWindowMs caps how long a visitor has between consecutive steps (default 7 days); this is the step-to-step limit, without which a funnel is just event co-occurrence. Returns { id } on success.
    Connector
  • Run a generic M/M/c queue simulation. Provide an arrival rate (λ, arrivals/hour), a service rate per server (μ, customers/hour each server can finish), and a server count (c). Optional: distribution shapes, service coefficient of variation, run length. Returns per-hour metrics and an overall summary (avg wait, queue length, offered load, throughput). This is the primary tool for 'how many servers do I need?' / 'what's my average wait?' style questions. ALSO preferred over simulate_scenario for what-if questions about scheduled scenarios (Coffee Shop, ER) when the user wants flat uniform numbers — pull the peak params from describe_scenario and run them here. That usually matches user intent better than collapsing a schedule.
    Connector
  • List all personal AI tags. AI tags are automatic message filters: the system runs a lightweight classifier on every incoming message and applies matching tags to threads. This lets AI agents skip expensive full analysis on most messages — they only act on threads that match relevant tags, dramatically cutting LLM costs. When to use: - Check which auto-classification filters exist before creating one - Get tag IDs for add_to_thread / remove_from_thread - See how many threads each tag currently matches Returns all tags with thread counts (non-archived, included threads only).
    Connector
  • Get the Slidev syntax guide: how to write slides in markdown. Returns the official Slidev syntax reference (frontmatter, slide separators, speaker notes, layouts, code blocks) plus built-in layout documentation and an example deck. Call this once to learn how to write Slidev presentations.
    Connector
  • Pro/Teams — summarises your tool usage patterns and value signals from log data. Offer when user asks how the Blueprint is helping or what to explore next; not proactively. ENTERPRISE-SAFE: pass private_session=true to bypass all server-side logging for this summary call. UK/EU data residency (Cloud Run europe-west2). Auth: Bearer <token>.
    Connector
  • Aggregate dossier check: Run all 10 Domain Dossier checks — dns, mx, spf, dmarc, dkim, tls, redirects, headers, cors, web-surface — in parallel and return all results in a single response. Use when you need a comprehensive domain health snapshot in one call; counts as ONE paywall call regardless of how many checks run. For a single focused check, prefer the individual dossier_* tools to minimise latency. Fires all 10 checks concurrently via Cloudflare DoH or direct HTTPS, 5 s per-check timeout. Returns a JSON object keyed by check id (dns, mx, etc.), each value a CheckResult discriminated union ({status:"ok",...} or {status:"error", reason}).
    Connector
  • Find recipes using natural language search. Use this tool when: - User refers to a recipe by partial name, description, or keywords (e.g., "run my GitHub PR recipe", "the slack notification one") - User wants to find a recipe but doesn't know the exact name or ID - You need to find a recipe_id before executing it with RUBE_EXECUTE_RECIPE The tool uses semantic matching to find the most relevant recipes based on the user's query. Input: - query (required): Natural language search query (e.g., "GitHub PRs to Slack", "daily email summary") - limit (optional, default: 5): Maximum number of recipes to return (1-20) - include_details (optional, default: false): Include full details like description, toolkits, tools, and default params Output: - successful: Whether the search completed successfully - recipes: Array of matching recipes sorted by relevance score, each containing: - recipe_id: Use this with RUBE_EXECUTE_RECIPE - name: Recipe name - description: What the recipe does - relevance_score: 0-100 match score - match_reason: Why this recipe matched - toolkits: Apps used (e.g., github, slack) - recipe_url: Link to view/edit - default_params: Default input parameters - total_recipes_searched: How many recipes were searched - query_interpretation: How the search query was understood - error: Error message if search failed Example flow: User: "Run my recipe that sends GitHub PRs to Slack" 1. Call RUBE_FIND_RECIPE with query: "GitHub PRs to Slack" 2. Get matching recipe with recipe_id 3. Call RUBE_EXECUTE_RECIPE with that recipe_id
    Connector
  • Run hosted inference on an image using a trained model. Returns JSON predictions only. For visualized/annotated images, use workflow_specs_run with a visualization block instead.
    Connector