Everything you need to know about markdown.new.
What is markdown.new?
It converts any public URL into clean, structured Markdown optimized for LLMs and AI agents. It strips away HTML bloat (ads, scripts, styling) while preserving headings, lists, links, and tables — making web content AI-ready with up to 80% fewer tokens.
How do I use it?
Three ways, no signup needed. Browser: prepend https://markdown.new/ to any URL. cURL/API: POST a JSON body with {"url": "https://example.com"}. Options: add method (auto, ai, browser) and retain_images (true/false) as query params or in the POST body.
Is it free?
Yes, always. No subscriptions, no paywalls. It’s an open utility for developers, AI builders, and teams. Just respect the fair usage limits so the service stays reliable for everyone.
What are the rate limits?
500 requests per day per IP address. If you exceed the limit, you’ll receive an HTTP 429 response. Check the x-rate-limit-remaining header to track your usage. For heavier workloads, consider self-hosting the open-source logic.
How does the conversion pipeline work?
It uses a three-tier fallback chain. Tier 1: requests the URL with Accept: text/markdown for native Markdown. Tier 2: passes HTML through Cloudflare Workers AI toMarkdown(). Tier 3: renders JS-heavy pages in a headless browser via Cloudflare Browser Rendering. The fastest successful tier wins.
Are there any limitations?
Public URLs only — paywalled or authenticated pages won’t work. JS-heavy sites using the browser method add ~1–2s of latency. Images are excluded by default (enable with retain_images=true). Very large pages may be truncated.
Is it legal to convert web pages?
Yes, for public pages. markdown.new uses standard HTTP requests (like your browser) and doesn’t store or republish content. Always respect each site’s Terms of Service and robots.txt. Avoid mass-scraping copyrighted content for redistribution.
Can I crawl an entire website?
Yes! Use
/crawl to crawl up to 100 pages from any site. It uses Cloudflare’s Browser Rendering
/crawl API with async job tracking. You can configure page limits (up to 100), crawl depth (up to 10), and enable JS rendering. Results are stored for 14 days.
How can I block this service from my site?
Our bot identifies as markdown.new/1.0. Add this to your robots.txt: User-agent: markdown.new Disallow: /. You can also block by User-Agent at the WAF/server level, use a llms.txt file to signal AI usage preferences, or return X-Robots-Tag: noindex headers.