Skip to main content
Glama

Firecrawl Local MCP Server

An MCP (Model Context Protocol) server for interacting with a self-hosted Firecrawl instance. This server provides web scraping and crawling capabilities through your local Firecrawl deployment.

Features

  • Web Scraping: Extract content from single web pages in markdown format

  • Web Crawling: Crawl entire websites with customizable depth and filtering

  • Site Mapping: Generate lists of all accessible URLs on a website

  • Job Monitoring: Track the status of crawling jobs

  • No API Key Required: Works directly with self-hosted Firecrawl instances

Installation

npm install npm run build

Configuration

The server connects to your Firecrawl instance using the FIRECRAWL_URL environment variable. By default, it connects to http://localhost:3002.

To change the Firecrawl URL, set the FIRECRAWL_URL environment variable in your MCP configuration.

Usage

With Claude Desktop

Add this to your Claude Desktop configuration file (claude_desktop_config.json):

{ "mcpServers": { "firecrawl-local": { "command": "node", "args": ["/absolute/path/to/firecrawl-local-mcp/dist/index.js"], "env": { "FIRECRAWL_URL": "http://localhost:3002" } } } }

With Cline

Add this to your Cline MCP configuration file:

{ "mcpServers": { "firecrawl-local": { "command": "node", "args": ["dist/index.js"], "cwd": "/absolute/path/to/firecrawl-local-mcp", "env": { "FIRECRAWL_URL": "http://localhost:3002" } } } }

Available Tools

firecrawl_scrape

Scrape a single webpage and return its content in markdown format.

Parameters:

  • url (required): The URL to scrape

  • formats: Output formats (default: ["markdown"])

  • onlyMainContent: Extract only main content (default: true)

  • includeTags: HTML tags to include

  • excludeTags: HTML tags to exclude

firecrawl_crawl

Crawl a website starting from a URL and return content from multiple pages.

Parameters:

  • url (required): The starting URL to crawl

  • includes: URL patterns to include (supports wildcards)

  • excludes: URL patterns to exclude (supports wildcards)

  • maxDepth: Maximum crawl depth (default: 2)

  • limit: Maximum number of pages to crawl (default: 10)

  • allowBackwardLinks: Allow crawling backward links (default: false)

  • allowExternalLinks: Allow crawling external links (default: false)

firecrawl_crawl_status

Check the status of a crawl job.

Parameters:

  • jobId (required): The job ID returned from a crawl request

firecrawl_map

Map a website to get a list of all accessible URLs.

Parameters:

  • url (required): The URL to map

  • search: Search query to filter URLs

  • ignoreSitemap: Ignore the website's sitemap (default: false)

  • includeSubdomains: Include subdomains (default: false)

  • limit: Maximum number of URLs to return (default: 5000)

Testing

Test the server functionality:

node test.js

This will test both the tool listing and a sample scrape operation.

Example Usage

Once configured in Claude Desktop, you can use natural language commands like:

  • "Scrape the content from https://example.com"

  • "Crawl the documentation site at https://docs.example.com with a depth of 3"

  • "Map all the URLs on https://example.com"

  • "Check the status of crawl job abc123"

Requirements

Troubleshooting

  1. Connection Issues: Verify your Firecrawl instance is running and accessible

  2. Timeout Errors: Adjust timeout values in src/index.ts for slow websites

  3. Authentication Errors: Ensure USE_DB_AUTHENTICATION=false in your Firecrawl .env file

-
security - not tested
F
license - not found
-
quality - not tested

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/ViperBlackSkull/firecrawl-local-mcp'

If you have feedback or need assistance with the MCP directory API, please join our Discord server