The WebSearch server enables you to:
Search: Perform advanced web searches to retrieve up-to-date information
Extract: Obtain specific information from web pages based on natural language prompts, with options to supplement with web searches and include sources
Crawl: Navigate through websites from a starting URL, with configurable depth and page limits
Scrape: Retrieve the full content from specified web pages
Uses .ENV files to manage API keys and configuration settings for the WebSearch server.
Provides integration for repository management, allowing users to fork, clone, and contribute to the WebSearch project.
Supports output of crawled and extracted web content in Markdown format for improved readability and structure.
Leverages OpenAI capabilities for enhanced features in web search and content analysis, requiring an API key for AI-powered functionality.
WebSearch - Advanced Web Search and Content Extraction Tool
A powerful web search and content extraction tool built with Python, leveraging the Firecrawl API for advanced web scraping, searching, and content analysis capabilities.
🚀 Features
Advanced Web Search: Perform intelligent web searches with customizable parameters
Content Extraction: Extract specific information from web pages using natural language prompts
Web Crawling: Crawl websites with configurable depth and limits
Web Scraping: Scrape web pages with support for various output formats
MCP Integration: Built as a Model Context Protocol (MCP) server for seamless integration
📋 Prerequisites
Python 3.8 or higher
uv package manager
Firecrawl API key
OpenAI API key (optional, for enhanced features)
Tavily API key (optional, for additional search capabilities)
🛠️ Installation
Install uv:
Clone the repository:
Create and activate a virtual environment with uv:
Install dependencies with uv:
Set up environment variables:
🎯 Usage
Setting Up With Claude for Desktop
Instead of running the server directly, you can configure Claude for Desktop to access the WebSearch tools:
Locate or create your Claude for Desktop configuration file:
Windows:
%env:AppData%\Claude\claude_desktop_config.json
macOS:
~/Library/Application Support/Claude/claude_desktop_config.json
Add the WebSearch server configuration to the
mcpServers
section:
Make sure to replace the directory path with the absolute path to your WebSearch project folder.
Save the configuration file and restart Claude for Desktop.
Once configured, the WebSearch tools will appear in the tools menu (hammer icon) in Claude for Desktop.
Available Tools
Search
Extract Information
Crawl Websites
Scrape Content
📚 API Reference
Search
query
(str): The search queryReturns: Search results in JSON format
Extract
urls
(List[str]): List of URLs to extract information fromprompt
(str): Instructions for extractionenableWebSearch
(bool): Enable supplementary web searchshowSources
(bool): Include source referencesReturns: Extracted information in specified format
Crawl
url
(str): Starting URLmaxDepth
(int): Maximum crawl depthlimit
(int): Maximum pages to crawlReturns: Crawled content in markdown/HTML format
Scrape
url
(str): Target URLReturns: Scraped content with optional screenshots
🔧 Configuration
Environment Variables
The tool requires certain API keys to function. We provide a .env.example
file that you can use as a template:
Copy the example file:
Edit the
.env
file with your API keys:
Getting the API Keys
OpenAI API Key:
Visit OpenAI's platform
Sign up or log in
Navigate to API keys section
Create a new secret key
Firecrawl API Key:
Visit Firecrawl's website
Create an account
Navigate to your dashboard
Generate a new API key
If everything is configured correctly, you should receive a JSON response with search results.
Troubleshooting
If you encounter errors:
Ensure all required API keys are set in your
.env
fileVerify the API keys are valid and have not expired
Check that the
.env
file is in the root directory of the projectMake sure the environment variables are being loaded correctly
🤝 Contributing
Fork the repository
Create your feature branch (
git checkout -b feature/AmazingFeature
)Commit your changes (
git commit -m 'Add some AmazingFeature'
)Push to the branch (
git push origin feature/AmazingFeature
)Open a Pull Request
📝 License
This project is licensed under the MIT License - see the LICENSE file for details.
🙏 Acknowledgments
Firecrawl for their powerful web scraping API
OpenAI for AI capabilities
MCPThe MCP community for the protocol specification
📬 Contact
José Martín Rodriguez Mortaloni - @m4s1t425 - jmrodriguezm13@gmail.com
Made with ❤️ using Python and Firecrawl
hybrid server
The server is able to function both locally and remotely, depending on the configuration or use case.
Built as a Model Context Protocol (MCP) server that provides advanced web search, content extraction, web crawling, and scraping capabilities using the Firecrawl API.
- 🚀 Features
- 📋 Prerequisites
- 🛠️ Installation
- 🎯 Usage
- 📚 API Reference
- 🔧 Configuration
- 🤝 Contributing
- 📝 License
- 🙏 Acknowledgments
- 📬 Contact
Related Resources
Related MCP Servers
- AsecurityAlicenseAqualityA Model Context Protocol (MCP) server that provides search and crawl functionality using Search1API.Last updated -126155MIT License
- AsecurityAlicenseAqualityA Model Context Protocol (MCP) server implementation that integrates with FireCrawl for advanced web scraping capabilities.Last updated -72,3284,531MIT License
- AsecurityAlicenseAqualityA Model Context Protocol server that enables web search, scraping, crawling, and content extraction through multiple engines including SearXNG, Firecrawl, and Tavily.Last updated -47455MIT License
- -securityAlicense-qualityA Model Context Protocol server that enables AI assistants to perform advanced web scraping, crawling, searching, and data extraction through the Firecrawl API.Last updated -72,328MIT License