Skip to main content
Glama

extract_urls

Crawl a webpage to extract, validate, and optionally queue all hyperlinks for processing, aiding in documentation discovery and graph building.

Instructions

Extract and analyze all URLs from a given web page. This tool crawls the specified webpage, identifies all hyperlinks, and optionally adds them to the processing queue. Useful for discovering related documentation pages, API references, or building a documentation graph. Handles various URL formats and validates links before extraction.

Input Schema

NameRequiredDescriptionDefault
add_to_queueNoIf true, automatically add extracted URLs to the processing queue for later indexing. This enables recursive documentation discovery. Use with caution on large sites to avoid excessive queuing.
urlYesThe complete URL of the webpage to analyze (must include protocol, e.g., https://). The page must be publicly accessible.

Input Schema (JSON Schema)

{ "properties": { "add_to_queue": { "default": false, "description": "If true, automatically add extracted URLs to the processing queue for later indexing. This enables recursive documentation discovery. Use with caution on large sites to avoid excessive queuing.", "type": "boolean" }, "url": { "description": "The complete URL of the webpage to analyze (must include protocol, e.g., https://). The page must be publicly accessible.", "type": "string" } }, "required": [ "url" ], "type": "object" }
Install Server

Other Tools from mcp-ragdocs

Related Tools

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/hannesrudolph/mcp-ragdocs'

If you have feedback or need assistance with the MCP directory API, please join our Discord server