The Olostep MCP server provides web scraping, content extraction, and structured Google search capabilities through integration with the Olostep platform.
Core Tools:
Web Content Extraction (
get_webpage_content) - Retrieves webpage content in clean markdown format with JavaScript rendering support and configurable wait times for dynamic sitesWebsite URL Discovery (
get_website_urls) - Discovers and maps URLs from any website, sorted by relevance to a search queryGoogle Search (
google_search) - Executes searches with structured data including organic results, knowledge graph information, People Also Ask questions, related searches, and rich snippets
Key Features:
Geo-targeted requests through specific countries (US, CA, GB, etc.) for localized content and search results
Comprehensive error handling for API issues, authentication failures, network errors, and rate limits
Simple authentication via
OLOSTEP_API_KEYenvironment variableCompatible with Claude Desktop, Windsurf, Cursor, and Metorial
Provides structured data from Google search results through the google_search tool, including organic search results, knowledge graph data, and related questions
Supports waiting for JavaScript-heavy websites to render before scraping content, with configurable wait times
Retrieves webpage content in clean markdown format for easier consumption and formatting
Olostep MCP Server
A Model Context Protocol (MCP) server implementation that integrates with Olostep for web scraping, content extraction, and search capabilities. To set up Olostep MCP Server, you need to have an API key. You can get the API key by signing up on the Olostep website.
Features
Scrape website content in HTML, Markdown, JSON or Plain Text (with optional parsers)
Parser-based web search with structured results
AI Answers with citations and optional JSON-shaped outputs
Batch scraping of up to 10k URLs
Autonomous site crawling from a start URL
Website URL discovery and mapping (with include/exclude filters)
Country-specific request routing for geo-targeted content
Configurable wait times for JavaScript-heavy websites
Comprehensive error handling and reporting
Simple API key configuration
Installation
🐳 Running with Docker (Recommended)
The easiest way to run the Olostep MCP server:
Local-only Docker build (no Docker Hub required)
If the Docker Hub image isn’t available from your environment, you can build and run the image locally from this repository:
Local smoke test (initialize + tools/list)
This MCP server uses stdio transport. You can validate it starts and lists tools without needing a working API key:
On Windows (PowerShell):
To actually call tools successfully, provide OLOSTEP_API_KEY when running the container.
Using Docker with Claude Desktop
Add this to your claude_desktop_config.json:
Using Docker with Cursor
Add an MCP server with:
Name:
olostepType:
commandCommand:
docker run -i --rm -e OLOSTEP_API_KEY=your-api-key olostep/mcp-server
Running with npx
On Windows (PowerShell):
On Windows (CMD):
Manual Installation
Running on Claude Desktop
Add this to your claude_desktop_config.json:
Or for a more straightforward way you can install via the Smithery CLI by running the following code in your device terminal
Running on Windsurf
Add this to your ./codeium/windsurf/model_config.json:
Running on Cursor
To configure Olostep MCP in Cursor:
Open Cursor Settings
Go to Features > MCP Servers
Click "+ Add New MCP Server"
Enter the following:
Name: "olostep-mcp" (or your preferred name)
Type: "command"
Command:
env OLOSTEP_API_KEY=your-api-key npx -y olostep-mcp
Replace your-api-key with your Olostep API key.
Running on Metorial
Option 1: One-Click Installation (Recommended)
Open Metorial dashboard
Navigate to MCP Servers directory
Search for "Olostep"
Click "Install" and enter your API key
Option 2: Manual Configuration
Add this to your Metorial MCP server configuration:
The Olostep tools will then be available in your Metorial AI chats.
Configuration
Environment Variables
OLOSTEP_API_KEY: Your Olostep API key (required)ORBIT_KEY: An optional key for using Orbit to route requests.
Available Tools
1. Scrape Website (scrape_website)
Extract content from a single URL. Supports multiple formats and JavaScript rendering.
Parameters:
url_to_scrape: The URL of the website you want to scrape (required)output_format: Choose format (html,markdown,json, ortext) - default:markdowncountry: Optional country code (e.g., US, GB, CA) for location-specific scrapingwait_before_scraping: Wait time in milliseconds before scraping (0-10000)parser: Optional parser ID for specialized extraction
Response (example):
2. Search the Web (search_web)
Search the Web for a given query and get structured results (non-AI, parser-based).
Parameters:
query: Search query (required)country: Optional country code for localized results (default:US)
Response:
Structured JSON (as text) representing parser-based results
3. Answers (AI) (answers)
Search the web and return AI-powered answers in the JSON structure you want, with sources and citations.
Parameters:
task: Question or task to answer using web data (required)json: Optional JSON schema/object or a short description of the desired output shape
Response includes:
answer_id,object,task,result(JSON if provided),sources,created
4. Batch Scrape URLs (batch_scrape_urls)
Scrape up to 10k URLs at the same time. Perfect for large-scale data extraction.
Response includes:
batch_id,status,total_urls,created_at,formats,country,parser,urls
5. Create Crawl (create_crawl)
Autonomously discover and scrape entire websites by following links.
Response includes:
crawl_id,object,status,start_url,max_pages,follow_links,created,formats,country,parser
6. Create Map (create_map)
Get all URLs on a website. Extract all URLs for discovery and analysis.
Response includes:
map_id,object,url,total_urls,urls,search_query,top_n
7. Get Webpage Content (get_webpage_content)
Retrieves webpage content in clean markdown format with support for JavaScript rendering.
Parameters:
url_to_scrape: The URL of the webpage to scrape (required)wait_before_scraping: Time to wait in milliseconds before starting the scrape (default: 0)country: Residential country to load the request from (e.g., US, CA, GB) (optional)
Response:
8. Get Website URLs (get_website_urls)
Search and retrieve relevant URLs from a website, sorted by relevance to your query.
Parameters:
url: The URL of the website to map (required)search_query: The search query to sort URLs by (required)
Response:
9. Web Search (google_search)
Retrieve structured data from web search results.
Parameters:
query: The search query to perform (required)country: Country code for localized results (e.g., US, GB) (default: "US")
Response includes:
Organic search results with titles, links, and snippets
Knowledge graph data when available
Related questions (People Also Ask)
Related searches
Rich snippets and other structured data
Error Handling
The server provides robust error handling:
Detailed error messages for API issues
Network error reporting
Authentication failure handling
Rate limit information
Example error response:
Distribution
Docker Images
The MCP server is available as a Docker image:
Docker Hub:
olostep/mcp-serverOfficial Docker MCP Registry:
mcp/olostep(coming soon - enhanced security with signatures & SBOMs)GitHub Container Registry:
ghcr.io/olostep/olostep-mcp-server
Docker Desktop MCP Toolkit
The Olostep MCP Server is being added to Docker Desktop's official MCP Toolkit, which means users will be able to:
Discover it in Docker Desktop's MCP Toolkit UI
Install it with one click
Configure it visually
Use it with any MCP-compatible client (Claude Desktop, Cursor, etc.)
Status: Submission in progress to Docker MCP Registry
Supported Platforms
linux/amd64linux/arm64
Building Locally
License
ISC License