Skip to main content
Glama

Scrapy MCP Server

by ThreeFish-AI

check_robots_txt

Analyze a website's robots.txt file to determine crawl permissions and ensure compliance with ethical web scraping practices. Provides insights into allowed and disallowed paths for crawling.

Instructions

Check the robots.txt file for a domain to understand crawling permissions.

This tool helps ensure ethical scraping by checking the robots.txt file of a website to see what crawling rules are in place.

Input Schema

NameRequiredDescriptionDefault
urlYes

Input Schema (JSON Schema)

{ "properties": { "url": { "title": "Url", "type": "string" } }, "required": [ "url" ], "type": "object" }

Other Tools from Scrapy MCP Server

Related Tools

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/ThreeFish-AI/scrapy-mcp'

If you have feedback or need assistance with the MCP directory API, please join our Discord server