How the crawl works
The crawler starts at your root URL (or sitemap), follows internal links breadth-first up to the depth you set, renders each page in a headless browser to capture JS-rendered content, runs the same readability + Markdown conversion as our single-page tool, and writes results to a folder mirroring the original URL paths. A typical 200-page docs site converts in 3-8 minutes; a 5000-page content library runs as a background job and emails you the ZIP when ready.
Single page or whole site?
If you have one URL in hand, use Webpage to Markdown — simpler workflow. Use Website to Markdown when you want to crawl across many pages: a docs site, a blog archive, a help center, a public knowledge base. Both produce identical per-page Markdown output; this tool just orchestrates the crawl, the deduplication, and the packaging.
Crawl politeness
By default crawls respect robots.txt, rate-limit to 1 request per second per host, and identify themselves with a clear User-Agent. All three are configurable for your own properties (where you control the rate limits) but stay polite by default for third-party sites. Don't use this to scrape sites whose terms of service forbid it; we don't bypass anything technical, but the social contract still applies.