Skip to main content
Glama

crawl

Extract website content by crawling from a seed URL, following links with configurable depth and page limits for structured data collection.

Instructions

Crawl a website starting from a seed URL, following links breadth-first up to a configurable depth and page limit.

Input Schema

TableJSON Schema
NameRequiredDescriptionDefault
concurrencyNoNumber of concurrent requests (default: 5)
depthNoMaximum link depth to follow (default: 2)
formatNoOutput format for each page: "markdown" (default), "llm", "text"
max_pagesNoMaximum number of pages to crawl (default: 50)
urlYesSeed URL to start crawling from
use_sitemapNoSeed the frontier from sitemap discovery before crawling

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/0xMassi/webclaw'

If you have feedback or need assistance with the MCP directory API, please join our Discord server