Skip to main content
Glama
NYO2008

Firecrawl MCP Server

by NYO2008

firecrawl_map

Discover all indexed URLs on a website to identify pages for scraping or locate specific sections. Returns an array of found URLs.

Instructions

Map a website to discover all indexed URLs on the site.

Best for: Discovering URLs on a website before deciding what to scrape; finding specific sections of a website. Not recommended for: When you already know which specific URL you need (use scrape or batch_scrape); when you need the content of the pages (use scrape after mapping). Common mistakes: Using crawl to discover URLs instead of map. Prompt Example: "List all URLs on example.com." Usage Example:

{
  "name": "firecrawl_map",
  "arguments": {
    "url": "https://example.com"
  }
}

Returns: Array of URLs found on the site.

Input Schema

TableJSON Schema
NameRequiredDescriptionDefault
urlYesStarting URL for URL discovery
searchNoOptional search term to filter URLs
ignoreSitemapNoSkip sitemap.xml discovery and only use HTML links
sitemapOnlyNoOnly use sitemap.xml for discovery, ignore HTML links
includeSubdomainsNoInclude URLs from subdomains in results
limitNoMaximum number of URLs to return

Implementation Reference

  • Handler for the firecrawl_map tool: validates arguments using isMapOptions, calls client.mapUrl with the provided URL and options, processes the response links, and returns them as text content.
    case 'firecrawl_map': {
      if (!isMapOptions(args)) {
        throw new Error('Invalid arguments for firecrawl_map');
      }
      const { url, ...options } = args;
      const response = await client.mapUrl(url, {
        ...options,
        // @ts-expect-error Extended API options including origin
        origin: 'mcp-server',
      });
      if ('error' in response) {
        throw new Error(response.error);
      }
      if (!response.links) {
        throw new Error('No links received from Firecrawl API');
      }
      return {
        content: [
          { type: 'text', text: trimResponseText(response.links.join('\n')) },
        ],
        isError: false,
      };
    }
  • Tool schema definition for firecrawl_map, including name, detailed description, and inputSchema specifying parameters like url, search, ignoreSitemap, etc.
    const MAP_TOOL: Tool = {
      name: 'firecrawl_map',
      description: `
    Map a website to discover all indexed URLs on the site.
    
    **Best for:** Discovering URLs on a website before deciding what to scrape; finding specific sections of a website.
    **Not recommended for:** When you already know which specific URL you need (use scrape or batch_scrape); when you need the content of the pages (use scrape after mapping).
    **Common mistakes:** Using crawl to discover URLs instead of map.
    **Prompt Example:** "List all URLs on example.com."
    **Usage Example:**
    \`\`\`json
    {
      "name": "firecrawl_map",
      "arguments": {
        "url": "https://example.com"
      }
    }
    \`\`\`
    **Returns:** Array of URLs found on the site.
    `,
      inputSchema: {
        type: 'object',
        properties: {
          url: {
            type: 'string',
            description: 'Starting URL for URL discovery',
          },
          search: {
            type: 'string',
            description: 'Optional search term to filter URLs',
          },
          ignoreSitemap: {
            type: 'boolean',
            description: 'Skip sitemap.xml discovery and only use HTML links',
          },
          sitemapOnly: {
            type: 'boolean',
            description: 'Only use sitemap.xml for discovery, ignore HTML links',
          },
          includeSubdomains: {
            type: 'boolean',
            description: 'Include URLs from subdomains in results',
          },
          limit: {
            type: 'number',
            description: 'Maximum number of URLs to return',
          },
        },
        required: ['url'],
      },
    };
  • src/index.ts:955-966 (registration)
    Registration of all tools including MAP_TOOL (firecrawl_map) in the ListToolsRequestSchema handler.
    server.setRequestHandler(ListToolsRequestSchema, async () => ({
      tools: [
        SCRAPE_TOOL,
        MAP_TOOL,
        CRAWL_TOOL,
        CHECK_CRAWL_STATUS_TOOL,
        SEARCH_TOOL,
        EXTRACT_TOOL,
        DEEP_RESEARCH_TOOL,
        GENERATE_LLMSTXT_TOOL,
      ],
    }));
  • Type guard helper function isMapOptions used to validate arguments for the firecrawl_map handler.
    function isMapOptions(args: unknown): args is MapParams & { url: string } {
      return (
        typeof args === 'object' &&
        args !== null &&
        'url' in args &&
        typeof (args as { url: unknown }).url === 'string'
      );
    }
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full burden and does well by explaining what the tool returns ('Array of URLs found on the site'), its discovery scope ('indexed URLs'), and behavioral constraints (e.g., it's for discovery, not for content). It could improve by mentioning potential limitations like rate limits or authentication needs.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is well-structured with clear sections (Best for, Not recommended for, Common mistakes, Prompt Example, Usage Example, Returns), front-loaded with key information, and every sentence adds value without redundancy.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness4/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the tool's complexity (6 parameters, no output schema, no annotations), the description is mostly complete, covering purpose, usage, examples, and return values. It could be slightly improved by adding more context on error handling or performance expectations, but it's largely sufficient.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, so the schema already documents all 6 parameters thoroughly. The description adds no additional parameter semantics beyond what's in the schema, but it meets the baseline of 3 since the schema does the heavy lifting.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool's purpose with specific verbs ('map', 'discover') and resources ('indexed URLs on the site'), and explicitly distinguishes it from sibling tools like 'scrape', 'batch_scrape', and 'crawl' by explaining what it's not for.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines5/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides explicit guidance with 'Best for' and 'Not recommended for' sections, naming specific alternatives (scrape, batch_scrape) and clarifying when not to use it (e.g., when you already know the URL). It also addresses common mistakes like using crawl instead of map.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

Install Server

Other Tools

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/NYO2008/firecrawl-mcp-server'

If you have feedback or need assistance with the MCP directory API, please join our Discord server