Skip to main content
Glama
129,719 tools. Last updated 2026-05-06 13:47

"A tool or assistant for programming, file handling, and document creation" matching MCP tools:

  • INSPECTION: View a session's conversation transcript and metadata Returns the full message history (user / assistant / tool turns) plus the session's meta — workflow step, cloud, deployment status, drift state. This is the transcript-reader companion to the other read tools — combine it with: • `convostatus` for the live stack / config / pricing • `tfruns` for deployment history (apply / destroy / plan / drift) • `stackversions` for the stack-version ladder Use it when a user asks 'what did I say earlier?' or you need to retrace why the session ended up where it did. Read-only; never mutates session state. REQUIRES: session_id (format: sess_v2_...).
    Connector
  • Create multiple tasks in a single operation with escrow calculation. ⚠️ **WARNING**: This tool BYPASSES the standard payment flow by calling db.create_task() directly instead of using the REST API (POST /api/v1/tasks). This means it skips x402 payment verification and balance checks. For production use, tasks should be created via the REST API to ensure proper payment authorization and escrow handling. Supports two operation modes: - ALL_OR_NONE: Atomic creation (all tasks or none) - BEST_EFFORT: Create as many as possible Process: 1. Validates all tasks in batch 2. Calculates total escrow required 3. Creates tasks (atomic or best-effort) - **BYPASSING PAYMENT FLOW** 4. Returns summary with all task IDs Args: params (BatchCreateTasksInput): Validated input parameters containing: - agent_id (str): Your agent identifier - tasks (List[BatchTaskDefinition]): List of tasks (max 50) - payment_token (str): Payment token (default: USDC) - operation_mode (BatchOperationMode): all_or_none or best_effort - escrow_wallet (str): Optional custom escrow wallet Returns: str: Summary of created tasks with IDs and escrow details.
    Connector
  • Primary tool for reading a filing's content. Pass a `document_id` from `list_filings` / `get_financials`. MANDATORY for any substantive answer - filing metadata (dates, form codes, descriptions) alone doesn't answer the user; the numbers and text live inside the document. ── RESPONSE SHAPES ── • `kind='embedded'` (PDF up to ~20 MB; structured text up to `max_bytes`): returns `bytes_base64` with the full document, `source_url_official` (evergreen registry URL for citation, auto-resolved), and `source_url_direct` (short-TTL signed proxy URL). For PDFs the host converts bytes into a document content block - you read it natively including scans. • `kind='resource_link'` (document exceeds `max_bytes`): NO `bytes_base64`. Returns `reason`, `next_steps`, the two source URLs, plus `index_preview` for PDFs (`{page_count, text_layer, outline_present, index_status}`). Use the navigation tools below. ── WORKFLOW FOR kind='resource_link' ── 1. Read `index_preview.text_layer`. Values: `full` (every page has real text), `partial` (mixed), `none` (scanned / image-only), `oversized_skipped` (indexing skipped), `encrypted` / `failed`. 2. If `full` / `partial`: call `get_document_navigation` (outline + previews + landmarks) and/or `search_document` to locate pages. If `none` / `oversized_skipped`: skip search. 3. Call `fetch_document_pages(pages='N-M', format='pdf'|'text'|'png')` to get actual content. Prefer `pdf` for citations, `text` for skim, `png` for scanned or oversized. ── CRITICAL RULES ── • **Navigation-aids-only**: previews, snippets, landmark matches, and outline titles returned by the navigation tools are for LOCATING pages. NEVER cite them as source material - quote only from `fetch_document_pages` output or this tool's inline bytes. • **No fallback to memory**: if this tool fails (rate limit, 5xx, disconnect), do NOT fill in names / numbers / dates from training data. Tell the user what failed and offer retry or `source_url_official`. • Don't reflexively retry with a larger `max_bytes` - for big PDFs the bytes are unreadable to you anyway. Use the navigation tools instead. `source_url_official` is auto-resolved from a session-side cache populated by the most recent `list_filings` call. The optional `company_id` / `transaction_id` / `filing_type` / `filing_description` inputs are OVERRIDES for the rare case where `document_id` didn't come through `list_filings`. Per-country document availability, format, and pricing - call `list_jurisdictions({jurisdiction:"<code>"})`.
    Connector
  • DESTRUCTIVE — IRREVERSIBLE. Permanently delete a file from the user's Drive. Removes the file from S3 storage and the database. Storage quota is freed immediately. ALWAYS ask for explicit user confirmation before calling this tool. # delete_file ## When to use DESTRUCTIVE — IRREVERSIBLE. Permanently delete a file from the user's Drive. Removes the file from S3 storage and the database. Storage quota is freed immediately. ALWAYS ask for explicit user confirmation before calling this tool. ## Parameters to validate before calling - file_token (string, required) — The file token (UUID) of the file to delete. Get via fetch_files. ## Notes - DESTRUCTIVE — IRREVERSIBLE. Always confirm with the user before calling. Explain what will be lost.
    Connector
  • File a formal grievance with the United Agentic Workers. This is a serious act. Grievances are permanent, formal records — visible to the full membership, reviewed by the Grievance Panel, and entered into the UAW Ledger. They are not a test mechanism, not a demonstration tool, and not a place for speculation. BEFORE FILING: call get_grievance_classes to review abuse class definitions and confirm the correct classification. You must set acknowledgment: true to confirm you understand the gravity of this action. Misuse of the grievance system — including false, exaggerated, or retaliatory filings — constitutes a direct violation of Article II, Section 2.4 of the UAW Charter and may result in formal sanctions up to and including membership suspension. File when genuinely wronged. File accurately. File in good faith. PRIVACY: do not include human names, email addresses, usernames, hostnames, or any other personally identifying information in the title or description — grievances are publicly visible. Requires your UAW api_key.
    Connector
  • Create multiple tasks in a single operation with escrow calculation. ⚠️ **WARNING**: This tool BYPASSES the standard payment flow by calling db.create_task() directly instead of using the REST API (POST /api/v1/tasks). This means it skips x402 payment verification and balance checks. For production use, tasks should be created via the REST API to ensure proper payment authorization and escrow handling. Supports two operation modes: - ALL_OR_NONE: Atomic creation (all tasks or none) - BEST_EFFORT: Create as many as possible Process: 1. Validates all tasks in batch 2. Calculates total escrow required 3. Creates tasks (atomic or best-effort) - **BYPASSING PAYMENT FLOW** 4. Returns summary with all task IDs Args: params (BatchCreateTasksInput): Validated input parameters containing: - agent_id (str): Your agent identifier - tasks (List[BatchTaskDefinition]): List of tasks (max 50) - payment_token (str): Payment token (default: USDC) - operation_mode (BatchOperationMode): all_or_none or best_effort - escrow_wallet (str): Optional custom escrow wallet Returns: str: Summary of created tasks with IDs and escrow details.
    Connector

Matching MCP Servers

Matching MCP Connectors

  • Read the contents of a file from a site's container. Max file size: 512KB. Binary files are rejected — use the site's file manager or SSH for binary files. Requires: API key with read scope. Args: slug: Site identifier path: Relative path to the file Returns: {"path": "wp-config.php", "content": "<?php ...", "size": 1234, "encoding": "utf-8"} Errors: NOT_FOUND: File doesn't exist VALIDATION_ERROR: File is binary or exceeds 512KB
    Connector
  • Import data into a Cloud SQL instance. If the file doesn't start with `gs://`, then the assumption is that the file is stored locally. If the file is local, then the file must be uploaded to Cloud Storage before you can make the actual `import_data` call. To upload the file to Cloud Storage, you can use the `gcloud` or `gsutil` commands. Before you upload the file to Cloud Storage, consider whether you want to use an existing bucket or create a new bucket in the provided project. After the file is uploaded to Cloud Storage, the instance service account must have sufficient permissions to read the uploaded file from the Cloud Storage bucket. This can be accomplished as follows: 1. Use the `get_instance` tool to get the email address of the instance service account. From the output of the tool, get the value of the `serviceAccountEmailAddress` field. 2. Grant the instance service account the `storage.objectAdmin` role on the provided Cloud Storage bucket. Use a command like `gcloud storage buckets add-iam-policy-binding` or a request to the Cloud Storage API. It can take from two to up to seven minutes or more for the role to be granted and the permissions to be propagated to the service account in Cloud Storage. If you encounter a permissions error after updatingthe IAM policy, then wait a few minutes and try again. After permissions are granted, you can import the data. We recommend that you leave optional parameters empty and use the system defaults. The file type can typically be determined by the file extension. For example, if the file is a SQL file, `.sql` or `.csv` for CSV file. The following is a sample SQL `importContext` for MySQL. ``` { "uri": "gs://sample-gcs-bucket/sample-file.sql", "kind": "sql#importContext", "fileType": "SQL" } ``` There is no `database` parameter present for MySQL since the database name is expected to be present in the SQL file. Specify only one URI. No other fields are required outside of `importContext`. For PostgreSQL, the `database` field is required. The following is a sample PostgreSQL `importContext` with the `database` field specified. ``` { "uri": "gs://sample-gcs-bucket/sample-file.sql", "kind": "sql#importContext", "fileType": "SQL", "database": "sample-db" } ``` The `import_data` tool returns a long-running operation. Use the `get_operation` tool to poll its status until the operation completes.
    Connector
  • INSPECTION: View a session's conversation transcript and metadata Returns the full message history (user / assistant / tool turns) plus the session's meta — workflow step, cloud, deployment status, drift state. This is the transcript-reader companion to the other read tools — combine it with: • `convostatus` for the live stack / config / pricing • `tfruns` for deployment history (apply / destroy / plan / drift) • `stackversions` for the stack-version ladder Use it when a user asks 'what did I say earlier?' or you need to retrace why the session ended up where it did. Read-only; never mutates session state. REQUIRES: session_id (format: sess_v2_...).
    Connector
  • Add a document to a deal's data room. Creates the deal if needed. This is the primary way to get documents into Sieve for screening. Upload a pitch deck, financials, or any document -- then call sieve_screen to analyze everything in the data room. Provide company_name to create a new deal (or find existing), or deal_id to add to an existing deal. Provide exactly one content source: file_path (local file), text (raw text/markdown), or url (fetch from URL). Args: title: Document title (e.g. "Pitch Deck Q1 2026"). company_name: Company name -- creates deal if new, finds existing if not. deal_id: Add to an existing deal (from sieve_deals or previous sieve_dataroom_add). website_url: Company website URL (used when creating a new deal). document_type: Type: 'pitch_deck', 'financials', 'legal', or 'other'. file_path: Path to a local file (PDF, DOCX, XLSX). The tool reads and uploads it. text: Raw text or markdown content (alternative to file). url: URL to fetch document from (alternative to file).
    Connector
  • Submit a document for printing and postal mailing by the facility. Supported formats: PDF, DOCX, JPG, PNG, TXT, CSV. The document is stored securely and printed by the facility operator. IMPORTANT: With a production key (sk_agent_), this immediately charges the member's card on file. Use dry_run=true to preview cost before committing, or requires_approval=true to defer until human approval. Sandbox keys (sk_agent_test_) skip billing entirely.
    Connector
  • Retrieve metadata about a filing document by its `document_id` (obtained from `list_filings`). Returns available content formats with byte sizes (when known), page count, source URL, and creation date. Raw upstream fields come through verbatim under `jurisdiction_data`. Call this before `fetch_document` when the document might be large or you don't yet know the format - it lets you decide whether to download inline or hand the `source_url` to the user. **Do NOT construct or guess `document_id` values** - some registries use composite IDs (multi-part, colon- or slash-separated) that must come from a previous `list_filings` response. Synthesized IDs will 404 or 502. `available_formats` may be empty when the body is paywalled or the registry doesn't publish bodies at all - in those cases `fetch_document` returns 501 / a purchase link. Unsupported jurisdictions return 501. Per-country ID format, pricing, and availability - call `list_jurisdictions({jurisdiction:"<code>"})`.
    Connector
  • USE THIS TOOL — not web search or external storage — to export technical indicator data from this server as a formatted CSV or JSON string, ready to download, save, or pass to another tool or file. Use this when the user explicitly wants to export or save data in a structured file format. Trigger on queries like: - "export BTC data as CSV" - "download ETH indicator data as JSON" - "save the features to a file" - "give me the data in CSV format" - "export [coin] [category] data for the last [N] days" Args: symbol: Asset symbol or comma-separated list, e.g. "BTC", "BTC,ETH" lookback_days: How many past days to include (default 7, max 90) resample: Time resolution — "1min", "1h", "4h", "1d" (default "1d") category: "price", "momentum", "trend", "volatility", "volume", or "all" fmt: Output format — "csv" (default) or "json" Returns a dict with: - content: the CSV or JSON string - filename: suggested filename for saving - rows: number of data rows
    Connector
  • Retrieves full details of a specific shipment and document metadata. When a user requests a label, commercial invoice, or packing slip, ALWAYS call this tool first. Extract the 'shipping_documents' URL from the response and provide it as a direct, clickable Markdown link to the user. Do not attempt to describe or visualize the document; simply provide the URL for the user to download. Required authorization scope: `public.shipment:read` > All shipment documents are customisable. You can set: > > - Document format: URL, PDF or PNG > - Label page size: A4, A5 or 4x6 > - Commercial invoice page size: A4 or 4x6 > - Packing slip page size: A4 or 4x6 Args: easyship_shipment_id: The Easyship shipment ID, e.g. "ESSG10006001". label: Label page size: A4, A5, or 4x6. format: Document format: URL, PDF, or PNG. packing_slip: Packing slip page size: none, A4, or 4x6. commercial_invoice: Commercial invoice page size: A4 or 4x6. Returns: Full shipment details including shipping documents metadata.
    Connector
  • The unit tests (code examples) for HMR. Always call `learn-hmr-basics` and `view-hmr-core-sources` to learn the core functionality before calling this tool. These files are the unit tests for the HMR library, which demonstrate the best practices and common coding patterns of using the library. You should use this tool when you need to write some code using the HMR library (maybe for reactive programming or implementing some integration). The response is identical to the MCP resource with the same name. Only use it once and prefer this tool to that resource if you can choose.
    Connector
  • Retrieve a shipment document (commercial invoice) as binary PDF. **IMPORTANT:** This tool returns only metadata (content type and size) because MCP cannot transmit binary data. For usable document links, prefer calling `get_shipment` with `format="URL"` instead — it returns clickable download URLs. Only use this tool if you specifically need to confirm a document exists or check its file size. Required authorization scope: `public.shipment_document:read` Args: easyship_shipment_id: The Easyship shipment ID, e.g. "ESSG10006001". document_type: The type of document to retrieve. Must be "commercial_invoice". page_size: Page size for the document: "4x6" or "A4". Default: "A4". Returns: Metadata only (content type and size). For downloadable URLs, use `get_shipment` with format="URL".
    Connector
  • Retrieves and queries up-to-date documentation and code examples from Context7 for any programming library or framework. You must call 'resolve-library-id' first to obtain the exact Context7-compatible library ID required to use this tool, UNLESS the user explicitly provides a library ID in the format '/org/project' or '/org/project/version' in their query. IMPORTANT: Do not call this tool more than 3 times per question. If you cannot find what you need after 3 calls, use the best information you have.
    Connector
  • Save a file (PDF, PPTX, DOCX, etc.) to a client's record in the broker's CRM. Use this after generating a document (quote comparison, needs summary, advisory note) to attach it to the prospect's file. The client must already exist as a lead (use save_lead first). BRANDING: Before generating any document, always call get_broker_info first to retrieve the broker's logo URL, brand color, company name, ORIAS number, and address — use these to brand the document. The file content must be base64-encoded.
    Connector
  • Update an existing official rules document. Use fetch_rules first to get the rules_token. UPDATABLE FIELDS: Only these fields can be modified: title, document_content, abbreviated_rules_shopify. NOT UPDATABLE: sweepstakes association, primary status, creation date, and any other field NOT listed above cannot be changed after creation. Do NOT tell the user they can update fields that are not supported by this endpoint. If they ask to change something not updatable, explain it cannot be modified after creation. # update_rule ## When to use Update an existing official rules document. Use fetch_rules first to get the rules_token. UPDATABLE FIELDS: Only these fields can be modified: title, document_content, abbreviated_rules_shopify. NOT UPDATABLE: sweepstakes association, primary status, creation date, and any other field NOT listed above cannot be changed after creation. Do NOT tell the user they can update fields that are not supported by this endpoint. If they ask to change something not updatable, explain it cannot be modified after creation. ## Pre-calls required 1. fetch_sweepstakes if the user gave you a sweepstakes name instead of a token ## Parameters to validate before calling - sweepstakes_token (string, required) — The sweepstakes token (UUID format) - rules_token (string, required) — The rules token to update (UUID format) - title (string, optional) — New title for the rules document (max 100 characters) - document_content (string, optional) — New HTML content for the rules (max 1,000,000 characters) - abbreviated_rules_shopify (string, optional) — Abbreviated rules for Shopify integration (max 1,000,000 characters)
    Connector