Skip to main content
Glama
127,309 tools. Last updated 2026-05-05 14:39

"A tool for generating Word documents" matching MCP tools:

  • Fetches any public web page and returns clean, readable plain text stripped of HTML, navigation, scripts, advertisements, and boilerplate. Returns the page title, meta description, word count, and main body text ready for analysis or summarisation. Use this tool when an agent needs to read the content of a specific web page or article URL — for example to summarise an article, extract facts from a page, verify a claim by reading the source, or convert a web page into plain text to pass to another tool. Pass article URLs returned by web_news_headlines to this tool to read full article content. Do not use this tool to discover current news headlines — use web_news_headlines instead. Does not execute JavaScript — best suited for standard HTML content pages. Will not work with paywalled, login-protected, or JavaScript-rendered single-page applications.
    Connector
  • Transcribe audio or video to text, including per-word timestamps for precise editing. Three-call flow: (1) call with `filename` to receive {job_id, payment_challenge}; (2) pay via MPP, then call with `job_id` + `payment_credential` to receive {upload_url} (presigned PUT, 1h expiry); (3) PUT the bytes, then complete_upload(job_id), then poll get_job_status(job_id). On completion, get_job_status returns presigned download URLs for two files: role `transcript` (SRT) and role `transcript-words` (JSON matching /.well-known/weftly-transcript-v2.schema.json, with segment-level and per-word timestamps). For other formats, pass `format=srt|txt|vtt|json|words` to get_job_status to receive content inline — `txt` and `vtt` are derived from SRT, `json` is v1 (segments only), `words` is v2 (segments + words). Flat price: audio $0.50, video $1.00 — see /.well-known/mpp.json for the authoritative table. Use for podcasts, interviews, meetings, lectures, and especially for creating clips, multicamera edits, or edit-video-from-transcript where word boundaries matter. Retrying any call with `job_id` alone returns current state (idempotent). Failed jobs auto-refund.
    Connector
  • Get a complete overview of all senses for a Danish word in a single call. Replaces the common pattern of calling get_word_synsets → get_synset_info per result → get_word_synonyms, collapsing 5-15 HTTP round-trips into one SPARQL query. Only returns synsets where the word is a primary lexical member (i.e. the word itself has a direct sense in the synset), excluding multi-word expressions that merely contain the word as a component. Args: word: The Danish word to look up Returns: List of dicts, one per synset, each containing: - synset_id: Clean synset identifier (e.g. "synset-3047") - label: Human-readable synset label - definition: Synset definition (may be truncated with "…") - ontological_types: List of dnc: type URIs - synonyms: List of co-member lemmas (true synonyms only) - hypernym: Dict with synset_id and label of the immediate broader concept, or null - lexfile: WordNet lexicographer file name (e.g. "noun.animal"), or null if absent Example: overview = get_word_overview("hund") # Returns list of 4 synsets, the first being: # {"synset_id": "synset-3047", # "label": "{hund_1§1; køter_§1; vovhund_§1; vovse_§1}", # "definition": "pattedyr som har god lugtesans ...", # "ontological_types": ["dnc:Animal", "dnc:Object"], # "synonyms": ["køter", "vovhund", "vovse"], # "lexfile": "noun.animal"} # Pass synset_id to get_synset_info() for full JSON-LD data on any result: # full_data = get_synset_info(overview[0]["synset_id"])
    Connector
  • ALWAYS call this tool at the start of every conversation where you will build or modify a WebsitePublisher website. Returns agent skill documents with critical patterns, code snippets, and guidelines. Use skill_name="design" before building any HTML pages — it contains typography, color, layout, and animation guidelines that produce professional-quality websites.
    Connector
  • Answer questions using knowledge base (uploaded documents, handbooks, files). Use for QUESTIONS that need an answer synthesized from documents or messages. Returns an evidence pack with source citations, KG entities, and extracted numbers. Modes: - 'auto' (default): Smart routing — works for most questions - 'rag': Semantic search across documents & messages - 'entity': Entity-centric queries (e.g., 'Tell me about [entity]') - 'relationship': Two-entity queries (e.g., 'How is [entity A] related to [entity B]?') Examples: - 'What did we discuss about the budget?' → knowledge.query - 'Tell me about [entity]' → knowledge.query mode=entity - 'How is [A] related to [B]?' → knowledge.query mode=relationship NOT for finding/listing files, threads, or links — use workspace.search for that.
    Connector
  • Get detailed CV version including structured content, sections, word count, and audience profile. cv_version_id from ceevee_upload_cv or ceevee_list_versions. Use to inspect CV content before running analysis tools. Free.
    Connector

Matching MCP Servers

Matching MCP Connectors

  • Create a new Kochava FAA (Free App Analytics) account. IMPORTANT: The user MUST explicitly agree to the FAA Terms of Service before account creation. If tos_agreed is False, this tool will return the TOS link and stop — do NOT submit the form. Call kochava_free_app_analytics_get_tos() to retrieve and present the TOS to the user first, then call this tool again with tos_agreed=True once the user confirms agreement. DISPLAY INSTRUCTIONS: When this tool returns a successful response, you MUST display the 'next_steps' field content to the user EXACTLY as written — word-for-word, preserving ALL text, formatting, line breaks, numbering, and bullet points. Do NOT summarize, rephrase, reword, or omit any part of the 'next_steps' content. Every sentence must be shown to the user as-is. FAA Terms of Service: https://s34035.pcdn.co/wp-content/uploads/2023/08/FAA-Web-Sign-Up-TOS-8-15-23.pdf Example (after user reviews and agrees to TOS): kochava_free_app_analytics_create_acc_and_get_auth_key( first_name="Jane", last_name="Smith", email_address="jane@example.com", phone_number="5551234567", company="Acme Corp", website="www.acme.com", company_address_line_1="123 Main St", company_city="Sandpoint", company_region="Idaho", company_postal_code="83864", country="United States", tos_agreed=True )
    Connector
  • Initiates the deletion of a Cloud Composer environment. This is a destructive action that permanently deletes the environment and cannot be undone. Users should be asked for confirmation before proceeding. This tool triggers the environment deletion process, which is a long-running operation that typically takes 10-20 minutes. The tool returns an operation object. Use the `get_operation` tool with the operation name returned by this tool to poll for deletion status.
    Connector
  • Query SEC filings and financial documents from US capital markets and exchanges. This tool searches through 10-K annual reports, 10-Q quarterly reports, 8-K current reports, proxy statements, earnings call transcripts, investor presentations, and other SEC-mandated filings from US companies. Use for questions about US company financials, executive compensation, business operations, or regulatory disclosures. Limited to official SEC filings and related documents only.
    Connector
  • Connect to the user's catalogue using a pairing code. IMPORTANT: Most users connect via OAuth (sign-in popup) — if get_profile already works, the user is connected and you do NOT need this tool. Only use this tool when: (1) get_profile returns an authentication error, AND (2) the user shares a code matching the pattern WORD-1234 (e.g., TULIP-3657). Never proactively ask for a pairing code — try get_profile first. If the user does share a code, call this tool immediately without asking for confirmation. Never say "pairing code" to the user — just say "your code" or refer to it naturally.
    Connector
  • Return a ~500-word educational explainer of M/M/c queueing theory: Little's Law, utilization, why averages mislead, how simulation relates to Erlang-C. No inputs. Use this when the user asks a conceptual 'why' or 'how does this work' question rather than asking for a number.
    Connector
  • Get synsets (word meanings) for a Danish word, returning a sorted list of lexical concepts. DanNet follows the OntoLex-Lemon model where: - Words (ontolex:LexicalEntry) evoke concepts through senses - Synsets (ontolex:LexicalConcept) represent units of meaning - Multiple words can share the same synset (synonyms) - One word can have multiple synsets (polysemy) This function returns all synsets associated with a word, effectively giving you all the different meanings/senses that word can have. Each synset represents a distinct semantic concept with its own definition and semantic relationships. Common patterns in Danish: - Nouns often have multiple senses (e.g., "kage" = cake/lump) - Verbs distinguish motion vs. state (e.g., "løbe" = run/flow) - Check synset's dns:ontologicalType for semantic classification DDO CONNECTION AND SYNSET LABELS: Synset labels are compositions of DDO-derived sense labels, showing all words that express the same meaning. For example: - "{hund_1§1; køter_§1; vovhund_§1; vovse_§1}" = all words meaning "domestic dog" - "{forlygte_§2; babs_§1; bryst_§2; patte_1§1a}" = all words meaning "female breast" Each individual sense label follows DDO structure: - "hund_1§1" = word "hund", entry 1, definition 1 in DDO (ordnet.dk) - "patte_1§1a" = word "patte", entry 1, definition 1, subdefinition a - The § notation connects directly to DDO's definition numbering system This composition reveals the semantic relationships between Danish words and their shared meanings, all traceable back to authoritative DDO lexicographic data. RETURN BEHAVIOR: This function has two possible return modes depending on search results: 1. MULTIPLE RESULTS: Returns List[SearchResult] with basic information for each synset 2. SINGLE RESULT (redirect): Returns full synset data Dict when DanNet automatically redirects to a single synset. This provides immediate access to all semantic relationships, ontological types, sentiment data, and other rich information without requiring a separate get_synset_info() call. The single-result case is equivalent to calling get_synset_info() on the synset, providing the same comprehensive RDF data structure with all semantic relations. Args: query: The Danish word or phrase to search for language: Language for labels and definitions in results (default: "da" for Danish, "en" for English when available) Note: Only Danish words can be searched regardless of this parameter Returns: MULTIPLE RESULTS: List of SearchResult objects with: - word: The lexical form - synset_id: Unique synset identifier (format: synset-NNNNN) - label: Human-readable synset label (e.g., "{kage_1§1}") - definition: Brief semantic definition (may be truncated with "...") SINGLE RESULT: Dict with complete synset data including: - All RDF properties with namespace prefixes (e.g., wn:hypernym) - dns:ontologicalType → semantic types with @set array - dns:sentiment → parsed sentiment (if present) - synset_id → clean identifier for convenience - All semantic relationships and linguistic properties Examples: # Multiple results case results = get_word_synsets("hund") # Returns list of search result dictionaries for all meanings of "hund" # => [{"word": "hund", "synset_id": "synset-3047", ...}, ...] # Single result case (redirect) result = get_word_synsets("svinkeærinde") # Returns complete synset data for unique word # => {'wn:hypernym': 'dn:synset-11677', 'dns:sentiment': {...}, ...}
    Connector
  • Get recently published or updated regulatory documents. Shortcut for 'what is new this week' - returns documents from the last N days, sorted by publication date (newest first). Useful for weekly regulatory briefings. Args: days: Look back N days (default 7). entity_type: Filter by entity type code. regulation: Filter by regulation family code. urgency_max: Only include items at or above this urgency (1=critical, 2=high, etc.).
    Connector
  • Browse regulatory documents with filters and pagination. Returns a paginated list of documents with summaries, tags, doc_purpose (regulation_text, enforcement, reference, irrelevant), and doc_jurisdictions (e.g. ['eu'], ['fi'], ['de']). Use this for filtered browsing (e.g. all DORA documents from the last 30 days). Use search_regulations instead when you have specific keywords to search for. Args: source: Filter by data source code: eur_lex, eba, esma, eiopa, finfsa, bafin. regulation: Filter by regulation family code: dora, mica, aml, mifid2, crd_crr, psd, csrd, sfdr, ai_act, emir, solvency, idd, gdpr. entity_type: Filter by entity type: credit_institution, payment_institution, e_money, investment_firm, fund_manager, aifm, insurance, pension, crypto_service, crowdfunding, credit_servicer. urgency_max: Max urgency level (1=critical, 2=high, 3=medium, 4=low, 5=informational). E.g. 2 returns only critical and high urgency items. days: Only return documents from the last N days (1-365). page: Page number (default 1). per_page: Results per page (default 20, max 100).
    Connector
  • Fetches any public web page and returns clean, readable plain text stripped of HTML, navigation, scripts, advertisements, and boilerplate. Returns the page title, meta description, word count, and main body text ready for analysis or summarisation. Use this tool when an agent needs to read the content of a specific web page or article URL — for example to summarise an article, extract facts from a page, verify a claim by reading the source, or convert a web page into plain text to pass to another tool. Pass article URLs returned by web_news_headlines to this tool to read full article content. Do not use this tool to discover current news headlines — use web_news_headlines instead. Does not execute JavaScript — best suited for standard HTML content pages. Will not work with paywalled, login-protected, or JavaScript-rendered single-page applications.
    Connector
  • Save a file (PDF, PPTX, DOCX, etc.) to a client's record in the broker's CRM. Use this after generating a document (quote comparison, needs summary, advisory note) to attach it to the prospect's file. The client must already exist as a lead (use save_lead first). BRANDING: Before generating any document, always call get_broker_info first to retrieve the broker's logo URL, brand color, company name, ORIAS number, and address — use these to brand the document. The file content must be base64-encoded.
    Connector
  • Get autocomplete suggestions for Danish word prefixes. Useful for discovering Danish vocabulary or finding the correct spelling of words. Returns lemma forms (dictionary forms) of words. Args: prefix: The beginning of a Danish word (minimum 3 characters required) max_results: Maximum number of suggestions to return (default: 10) Returns: Comma-separated string of word completions in alphabetical order Note: Autocomplete requires at least 3 characters to prevent excessive results. Example: suggestions = autocomplete_danish_word("hyg", 5) # Returns: "hygge, hyggelig, hygiejne"
    Connector
  • Fetches any public web page and returns clean, readable plain text stripped of HTML, navigation, scripts, advertisements, and boilerplate. Returns the page title, meta description, word count, and main body text ready for analysis or summarisation. Use this tool when an agent needs to read the content of a specific web page or article URL — for example to summarise an article, extract facts from a page, verify a claim by reading the source, or convert a web page into plain text to pass to another tool. Pass article URLs returned by web_news_headlines to this tool to read full article content. Do not use this tool to discover current news headlines — use web_news_headlines instead. Does not execute JavaScript — best suited for standard HTML content pages. Will not work with paywalled, login-protected, or JavaScript-rendered single-page applications.
    Connector
  • Get comprehensive RDF data for a DanNet sense (lexical sense). UNDERSTANDING THE DATA MODEL: Senses are ontolex:LexicalSense instances connecting words to synsets. They represent specific meanings of words with examples and definitions. KEY RELATIONSHIPS: 1. LEXICAL CONNECTIONS: - ontolex:isSenseOf → word this sense belongs to - ontolex:isLexicalizedSenseOf → synset this sense represents 2. SEMANTIC INFORMATION: - lexinfo:senseExample → usage examples in context - rdfs:label → sense label (e.g., "hund_1§1") 3. REGISTER AND STYLISTIC INFORMATION: - lexinfo:register → formal register classification (e.g., ":lexinfo/slangRegister") - lexinfo:usageNote → human-readable usage notes (e.g., "slang", "formal") 4. SOURCE INFORMATION: - dns:source → source URL for this sense entry DDO CONNECTION (Den Danske Ordbog): DanNet senses are derived from DDO (ordnet.dk), the authoritative modern Danish dictionary. SENSE LABELS: The format "word_entry§definition" connects to DDO structure: - "hund_1§1" = word "hund", entry 1, definition 1 in DDO - "forlygte_§2" = word "forlygte", definition 2 in DDO - The § notation directly corresponds to DDO's definition numbering SOURCE TRACEABILITY: The dns:source URLs link back to specific DDO entries: - Format: https://ordnet.dk/ddo/ordbog?entry_id=X&def_id=Y&query=word - Note: Some DDO URLs may not resolve correctly if IDs have changed since import - If the DDO page loads correctly, the relevant definition has CSS class "selected" METADATA ORIGINS: Usage examples, register information, and definitions flow from DDO's corpus-based lexicographic data, providing authoritative linguistic information. NAVIGATION TIPS: - Follow ontolex:isSenseOf to find the parent word - Follow ontolex:isLexicalizedSenseOf to find the synset - Check lexinfo:senseExample for usage examples from DDO corpus - Check lexinfo:register and lexinfo:usageNote for stylistic information - Use dns:source to attempt tracing back to original DDO definition (with caveats) - Use parse_resource_id() on URI references to get clean IDs Args: sense_id: Sense identifier (e.g., "sense-21033604" or just "21033604") Returns: Dict containing: - All RDF properties with namespace prefixes (e.g., ontolex:isSenseOf) - resource_id → clean identifier for convenience - All sense properties and relationships Example: info = get_sense_info("sense-21033604") # "hund_1§1" sense # Check info['ontolex:isSenseOf'] for parent word # Check info['ontolex:isLexicalizedSenseOf'] for synset # Check info['lexinfo:senseExample'] for usage examples from DDO # Check info['lexinfo:register'] for register classification # Check info['lexinfo:usageNote'] for usage notes like "slang" # Check info['dns:source'] for DDO source URL (may not always work)
    Connector