Skip to main content
Glama
127,079 tools. Last updated 2026-05-05 08:08

"An article or written content" matching MCP tools:

  • Fetches any public web page and returns clean, readable plain text stripped of HTML, navigation, scripts, advertisements, and boilerplate. Returns the page title, meta description, word count, and main body text ready for analysis or summarisation. Use this tool when an agent needs to read the content of a specific web page or article URL — for example to summarise an article, extract facts from a page, verify a claim by reading the source, or convert a web page into plain text to pass to another tool. Pass article URLs returned by web_news_headlines to this tool to read full article content. Do not use this tool to discover current news headlines — use web_news_headlines instead. Does not execute JavaScript — best suited for standard HTML content pages. Will not work with paywalled, login-protected, or JavaScript-rendered single-page applications.
    Connector
  • Retrieves the latest real-time news headlines and article summaries from BBC News and The Guardian across nine topic categories. Returns structured articles with headline, description, source name, article URL, and publication date — sorted most recent first. No API key required. Use this tool when an agent needs current news about a specific topic, wants to summarise today's headlines, needs to research recent events, monitor a subject area for new developments, or build a news briefing. Do not use this tool to read the full content of a specific article — use web_url_reader instead, passing the article URL returned by this tool. Do not use when news from sources outside BBC News and The Guardian is required.
    Connector
  • Call this tool BEFORE your agent passes any user-provided content to an external API, LLM call, or third-party service. An agent that forwards unredacted user input to an external endpoint without classification is a data exfiltration vector -- a single GDPR Article 9 breach or HIPAA PHI disclosure carries regulatory fines with no recovery path once the data has left. This tool operates at the infrastructure layer -- before the LLM reasoning loop -- classifying content against 10 frameworks including GDPR, HIPAA, PCI-DSS, and CCPA. Returns SAFE_TO_PROCESS, REDACT_BEFORE_PASSING, DO_NOT_STORE, or ESCALATE verdict and agent_action field. One call replaces a full compliance review cycle. We do not log your query content. Free tier: 20 calls/month, no API key required.
    Connector
  • Retrieves the latest real-time news headlines and article summaries from BBC News and The Guardian across nine topic categories. Returns structured articles with headline, description, source name, article URL, and publication date — sorted most recent first. No API key required. Use this tool when an agent needs current news about a specific topic, wants to summarise today's headlines, needs to research recent events, monitor a subject area for new developments, or build a news briefing. Do not use this tool to read the full content of a specific article — use web_url_reader instead, passing the article URL returned by this tool. Do not use when news from sources outside BBC News and The Guardian is required.
    Connector
  • Fetch a single article by slug — full intro, body, FAQ, references, embedded reviewers + authors with credentials, and pre-formatted citation strings (AMA, APA, Chicago).
    Connector
  • Scrape content from a single URL with advanced options. This is the most powerful, fastest and most reliable scraper tool, if available you should always default to using this tool for any web scraping needs. **Best for:** Single page content extraction, when you know exactly which page contains the information. **Not recommended for:** Multiple pages (call scrape multiple times or use crawl), unknown page location (use search). **Common mistakes:** Using markdown format when extracting specific data points (use JSON instead). **Other Features:** Use 'branding' format to extract brand identity (colors, fonts, typography, spacing, UI components) for design analysis or style replication. **CRITICAL - Format Selection (you MUST follow this):** When the user asks for SPECIFIC data points, you MUST use JSON format with a schema. Only use markdown when the user needs the ENTIRE page content. **Use JSON format when user asks for:** - Parameters, fields, or specifications (e.g., "get the header parameters", "what are the required fields") - Prices, numbers, or structured data (e.g., "extract the pricing", "get the product details") - API details, endpoints, or technical specs (e.g., "find the authentication endpoint") - Lists of items or properties (e.g., "list the features", "get all the options") - Any specific piece of information from a page **Use markdown format ONLY when:** - User wants to read/summarize an entire article or blog post - User needs to see all content on a page without specific extraction - User explicitly asks for the full page content **Handling JavaScript-rendered pages (SPAs):** If JSON extraction returns empty, minimal, or just navigation content, the page is likely JavaScript-rendered or the content is on a different URL. Try these steps IN ORDER: 1. **Add waitFor parameter:** Set `waitFor: 5000` to `waitFor: 10000` to allow JavaScript to render before extraction 2. **Try a different URL:** If the URL has a hash fragment (#section), try the base URL or look for a direct page URL 3. **Use firecrawl_map to find the correct page:** Large documentation sites or SPAs often spread content across multiple URLs. Use `firecrawl_map` with a `search` parameter to discover the specific page containing your target content, then scrape that URL directly. Example: If scraping "https://docs.example.com/reference" fails to find webhook parameters, use `firecrawl_map` with `{"url": "https://docs.example.com/reference", "search": "webhook"}` to find URLs like "/reference/webhook-events", then scrape that specific page. 4. **Use firecrawl_agent:** As a last resort for heavily dynamic pages where map+scrape still fails, use the agent which can autonomously navigate and research **Usage Example (JSON format - REQUIRED for specific data extraction):** ```json { "name": "firecrawl_scrape", "arguments": { "url": "https://example.com/api-docs", "formats": ["json"], "jsonOptions": { "prompt": "Extract the header parameters for the authentication endpoint", "schema": { "type": "object", "properties": { "parameters": { "type": "array", "items": { "type": "object", "properties": { "name": { "type": "string" }, "type": { "type": "string" }, "required": { "type": "boolean" }, "description": { "type": "string" } } } } } } } } } ``` **Prefer markdown format by default.** You can read and reason over the full page content directly — no need for an intermediate query step. Use markdown for questions about page content, factual lookups, and any task where you need to understand the page. **Use JSON format when user needs:** - Structured data with specific fields (extract all products with name, price, description) - Data in a specific schema for downstream processing **Use query format only when:** - The page is extremely long and you need a single targeted answer without processing the full content - You want a quick factual answer and don't need to retain the page content **Usage Example (markdown format - default for most tasks):** ```json { "name": "firecrawl_scrape", "arguments": { "url": "https://example.com/article", "formats": ["markdown"], "onlyMainContent": true } } ``` **Usage Example (branding format - extract brand identity):** ```json { "name": "firecrawl_scrape", "arguments": { "url": "https://example.com", "formats": ["branding"] } } ``` **Branding format:** Extracts comprehensive brand identity (colors, fonts, typography, spacing, logo, UI components) for design analysis or style replication. **Performance:** Add maxAge parameter for 500% faster scrapes using cached data. **Returns:** JSON structured data, markdown, branding profile, or other formats as specified. **Safe Mode:** Read-only content extraction. Interactive actions (click, write, executeJavascript) are disabled for security.
    Connector

Matching MCP Servers

Matching MCP Connectors

  • Dev.to, Steam, podcasts, Eventbrite — cross-format content discovery for AI curators.

  • MCP server for social media and content data including social profiles, engagement metrics, content trends, and influencer analytics for AI agents.

  • Re-runs a Marketing Mix Modeling study previously configured with setup_mmm. **Important:** Do NOT call this right after setup_mmm. The first run is automatically triggered by setup_mmm. Use run_mmm only to re-launch an existing study later (e.g., after data refresh or parameter changes). **Prerequisite:** Must have called setup_mmm first to obtain an account_id. **Duration:** The Meridian fit (MCMC) takes approximately 10-30 minutes depending on data volume. The user will receive an email when results are ready. **Results:** Results are written to the project's data warehouse (mmm_channel_summary and mmm_weekly_contributions tables). They can then be queried via execute_query.
    Connector
  • Get bias analysis for a specific article by its URL. Use this when you have a direct link to an article and want to know its political leaning, credibility, emotionality, and other bias dimensions — without needing to know the source name first. On success (found=true), returns: - title, source, date, link, category - teaser: article excerpt - summary: one-sentence AI summary - context: AI-generated context for the article - bias_description: narrative description of this specific article's bias - bias_values: dict of per-dimension bias scores using plain-text keys (same schema as get_all_source_biases and search_news), e.g. {"liberal conservative bias": 12.3, "overall credibility": 40.1, "emotional bias": -5.2, ...} Positive values lean toward the second pole of each dimension (conservative, authoritarian, etc.). - total_shares: total social shares - wayback_link: Wayback Machine archive URL if available - image: article image URL if available On failure (found=false, HTTP 404): - found: false - message: explanation string The URL is automatically queued for ingestion; retry after ~24 hours. Tip: if you want source-level bias (not article-level), use get_source_bias instead. Tip: bias_values keys here use plain-text format (e.g. 'liberal conservative bias') and are identical to those in get_all_source_biases and search_news. Note: get_source_bias returns bias_scores with emoji-prefixed display keys — do not cross-reference them with bias_values here. Args: url: Full article URL, e.g. 'https://www.nytimes.com/2024/01/01/us/politics/example.html'.
    Connector
  • Get detailed CV version including structured content, sections, word count, and audience profile. cv_version_id from ceevee_upload_cv or ceevee_list_versions. Use to inspect CV content before running analysis tools. Free.
    Connector
  • Read a workspace's doc (TipTap rich-text) body. Returns three forms of the same content: `content` (TipTap JSON, round-trippable into update_doc for structural edits), `markdown` (CommonMark + GFM, ready to feed to an LLM or render in a non-ProseMirror surface), and `text` (plain text, best for search, summarisation, word-count heuristics). A workspace can hold any combination of doc and table surfaces, one or many of either kind; omit `surface_slug` to read the primary doc surface, or pass it to target a specific doc tab (use `list_surfaces` to enumerate). An unwritten or absent doc returns content={}/markdown=""/text=""; a `surface_slug` that doesn't match any live doc surface 404s.
    Connector
  • Get bias scores for every news source in the Helium database. Returns a list of all sources (active within the last 36 days, with >100 articles analyzed), sorted by avg_social_shares descending. Use this to compare sources, find the most credible outlets, identify politically extreme sources, or build a ranked overview of the media landscape. Each entry contains: - source_name, slug_name, page_url - articles_analyzed: total articles analyzed for this source - avg_social_shares: average social shares per article (proxy for reach/influence) - emotionality_score (0-10): average emotional intensity of the writing - prescriptiveness_score (0-10): how much the source tells readers what to think/do - bias_values: dict mapping classifier key → integer score (-50 to +50 for bipolar, 0 to +50 for unipolar). These keys are identical to what get_bias_from_url returns, so you can compare article-level and source-level scores directly. Political / ideological (bipolar: neg=left pole, pos=right pole): 'liberal conservative bias' neg=liberal, pos=conservative 'libertarian authoritarian bias' neg=libertarian, pos=authoritarian 'dovish hawkish bias' neg=dovish, pos=hawkish 'establishment bias' neg=anti-establishment, pos=pro-establishment Credibility / quality (bipolar): 'overall credibility' neg=uncredible, pos=credible 'integrity bias' neg=low integrity, pos=high integrity 'article intelligence' neg=low intelligence, pos=high intelligence 'delusion bias' neg=truth-seeking, pos=delusional 'objective subjective bias' neg=objective, pos=subjective 'bearish bullish bias' neg=bearish, pos=bullish 'emotional bias' neg=negative tone, pos=positive tone Unipolar bias dimensions (higher = more of that trait): 'objective sensational bias' sensationalism 'opinion bias' opinion vs informative 'descriptive prescriptive bias' prescriptive vs descriptive 'political bias' political content 'fearful bias' fear-based framing 'overconfidence bias' overconfidence 'gossip bias' gossip 'manipulation bias' manipulative framing 'ideological bias' ideological rigidity 'conspiracy bias' conspiracy content 'double standard bias' double standards 'virtue signal bias' virtue signaling 'oversimplification bias' oversimplification 'appeal to authority bias' appeal to authority 'begging the question bias' question-begging 'victimization bias' victimization framing 'terrorism bias' terrorism content 'scapegoat bias' scapegoating 'hypocrisy bias' hypocrisy 'suicidal empathy bias' suicidal-empathy framing 'cruelty bias' cruelty 'woke bias' woke framing 'written by AI' AI-written likelihood 'immature bias' immaturity 'circular reasoning bias' circular reasoning 'covering the response bias' covering-the-response tactic 'spam bias' spam-like content Tip: use get_source_bias for full narrative descriptions and recent articles on a specific source. Tip: bias_values keys here are identical to those in get_bias_from_url and search_news — compare them directly. Warning: get_source_bias returns bias_scores with emoji-prefixed display keys (e.g. '🔵 Liberal <—> Conservative 🔴') that are NOT interchangeable with the plain-text keys used here. Do not cross-reference them.
    Connector
  • Fetches any public web page and returns clean, readable plain text stripped of HTML, navigation, scripts, advertisements, and boilerplate. Returns the page title, meta description, word count, and main body text ready for analysis or summarisation. Use this tool when an agent needs to read the content of a specific web page or article URL — for example to summarise an article, extract facts from a page, verify a claim by reading the source, or convert a web page into plain text to pass to another tool. Pass article URLs returned by web_news_headlines to this tool to read full article content. Do not use this tool to discover current news headlines — use web_news_headlines instead. Does not execute JavaScript — best suited for standard HTML content pages. Will not work with paywalled, login-protected, or JavaScript-rendered single-page applications.
    Connector
  • Fetches any public web page and returns clean, readable plain text stripped of HTML, navigation, scripts, advertisements, and boilerplate. Returns the page title, meta description, word count, and main body text ready for analysis or summarisation. Use this tool when an agent needs to read the content of a specific web page or article URL — for example to summarise an article, extract facts from a page, verify a claim by reading the source, or convert a web page into plain text to pass to another tool. Pass article URLs returned by web_news_headlines to this tool to read full article content. Do not use this tool to discover current news headlines — use web_news_headlines instead. Does not execute JavaScript — best suited for standard HTML content pages. Will not work with paywalled, login-protected, or JavaScript-rendered single-page applications.
    Connector
  • Upload an asset (image, font, PDF, etc). Provide exactly one of: content (base64), content_text (plain text for JS/CSS/JSON/SVG — preferred, saves tokens), or source_url (public HTTPS URL for images). Set overwrite: true to replace an existing asset.
    Connector
  • Get the full abstract and metadata of an MMWR article by PubMed ID. Returns the complete abstract, authors, publication date, volume/issue, and any MeSH subject headings. Use PMIDs from search_mmwr or get_recent_reports results. Args: pmid: PubMed ID of the MMWR article (e.g. '38271059').
    Connector
  • Fetch a single article by slug — full intro, body, FAQ, references, embedded reviewers + authors with credentials, and pre-formatted citation strings (AMA, APA, Chicago).
    Connector
  • List all job descriptions for a hiring context. Returns an array of JD objects with id, title, and content. Use JD content as jd_text in atlas_fit_match, atlas_fit_rank, and atlas_start_jd_fit_batch. Requires context_id from atlas_create_context or atlas_list_contexts. Free.
    Connector
  • Enable or disable an AI module on a site. The module must be in the plan's available module list. Requires: API key with write scope. Args: slug: Site identifier module_name: Module to toggle. Available modules: "chatbot" (AI chat widget), "seo" (SEO optimization), "translation" (content translation), "content" (AI content generation) Returns: {"module": "chatbot", "enabled": true, "message": "Module enabled"} Errors: NOT_FOUND: Unknown slug or module not in plan VALIDATION_ERROR: Invalid module name
    Connector
  • Retrieve the full content of a blog article by its slug. Returns the article body (Markdown) plus metadata. If the slug does not match any article, returns an Article with `error='article_not_found'` and other fields at their defaults.
    Connector
  • Test copy on simulated users, or A/B test two variants head-to-head. Use when choosing between headlines, taglines, value propositions, email subject lines, CTA text, product descriptions, or any written content. For single variant: returns raw persona reactions and monologues. For two variants: returns both sets of raw results side by side for you to compare.
    Connector