Skip to main content

MCP server that fetches URLs and returns clean Markdown, with Cloudflare fast-path and Chromium fallback

Project description

web-to-markdown-mcp

An MCP server that fetches a URL through a real Chromium browser and returns the main content as clean Markdown.

Why

Most MCP web-fetch tools either:

  • use plain HTTP, which fails on JS-required pages and gets blocked by Cloudflare bot detection on many sites; or
  • use a real browser but return raw HTML or accessibility-tree snapshots, which are noisy and token-heavy when you just want to read the article.

This server uses a two-tier approach:

  1. Native markdown fast-path. Every request first tries a plain HTTP GET with an Accept: text/markdown header. Servers that support content negotiation — such as Cloudflare-hosted sites with Markdown for Agents enabled — respond with Content-Type: text/markdown, and the body is returned immediately with no browser overhead. Servers that don't recognise the header respond normally and fall through to tier 2.

  2. Browser fallback. When the fast-path doesn't yield markdown, patchright (a Playwright fork with anti-detection patches) drives real Chromium, and trafilatura strips navigation, sidebars, ads, and footers down to the article body as clean Markdown.

After navigation, the server polls the DOM and runs trafilatura, returning as soon as two consecutive polls produce the same extraction. This means it returns within a few hundred milliseconds for typical pages — rather than waiting for analytics, ads, and other late-loading resources to finish — and gives slow SPAs and bot-challenge clearance time to settle without timing out prematurely.

Headless mode is the default and works on standard pages. For sites with active bot detection (Cloudflare challenges and similar), pass headless=False to use a visible Chromium window — slower and visually intrusive, but clears most challenges that block headless mode.

For a typical article, expect roughly 80% fewer tokens than the raw HTML and roughly 90% fewer than a full accessibility-tree snapshot.

Installation

Requires Python 3.10+ and a one-time Chromium download (~300 MB).

# Run directly with uv (no install step)
uvx web-to-markdown-mcp

# Or install with pip
pip install web-to-markdown-mcp

# One-time browser download
patchright install chromium

MCP client configuration

Claude Desktop

Edit your config file:

  • macOS: ~/Library/Application Support/Claude/claude_desktop_config.json
  • Windows: %APPDATA%\Claude\claude_desktop_config.json
{
  "mcpServers": {
    "web-to-markdown": {
      "command": "uvx",
      "args": ["web-to-markdown-mcp"]
    }
  }
}

Restart Claude Desktop.

LM Studio

Edit ~/.lmstudio/mcp.json (Developer tab → Edit mcp.json) — same JSON block as above. Then enable Allow calling servers from mcp.json in the Developer tab's Server Settings. The server appears in the Integrations tab of any new chat.

Cursor / Windsurf / VS Code

Same JSON block, in each client's MCP config location.

Claude Code

claude mcp add web-to-markdown -- uvx web-to-markdown-mcp

Usage

The server exposes a single tool:

fetch_url_as_markdown

Parameter Type Default Description
url string required The URL to fetch
wait_until string "domcontentloaded" When navigation completes: "load", "domcontentloaded", "networkidle", "commit"
timeout_ms int 60000 Navigation-step timeout in milliseconds
headless bool true false uses a visible browser window — slower but clears more bot detection
poll_budget_ms int 5000 Max time after navigation to wait for content stabilization
poll_interval_ms int 250 How often to re-attempt extraction during polling

Returns Markdown as a string, or a string beginning with "ERROR:" on expected failures (timeout, no extractable content, navigation error).

Example call (from any MCP client's tool-use UI):

fetch_url_as_markdown(url="https://example.com/long-article")

wait_until choice:

  • "domcontentloaded" (default) — returns when the DOM is built; content-stabilization polling handles the rest
  • "load" — waits for all subresources (images, scripts, stylesheets); rarely needed since polling runs after this
  • "networkidle" — waits for network to quiet; sometimes hangs on pages with persistent background connections
  • "commit" — returns as soon as the response starts; rarely useful

When to bump poll_budget_ms: the 5-second default is fine for typical pages but may return a partial extraction on slow SPAs that render content over many seconds, and may time out before a bot-detection challenge clears in headed mode. For headed-mode fetches of bot-protected sites, 10000-15000 is a reasonable budget.

Limitations

  • Cold start. Each call launches a fresh Chromium (~1–2 s overhead). A persistent browser instance is on the v2 roadmap.
  • Bot detection on hard sites. patchright headless clears default Cloudflare configurations on many sites, but pages running aggressive Bot Fight Mode, Turnstile interactive challenges, or commercial bot-management products (PerimeterX, DataDome, Kasada) — and Cloudflare's own marketing site — still detect headless Chromium. Pass headless=False to use a visible browser window, which clears most of these. The cost is a Chromium window flashing on screen for a couple of seconds per fetch.
  • Headed-mode bot challenges need a generous polling budget. When using headless=False on bot-protected sites, the challenge can take 5-15 seconds to clear. Set poll_budget_ms to 10000-15000 for these cases — the 5000 default may return prematurely while the challenge is still resolving.
  • Headed mode requires a display. headless=False fails on servers without a graphical environment (cloud VMs, containers, CI). Use a virtual display like Xvfb if you need headed mode in those environments.
  • Datacenter IPs. Cloudflare's harder challenges still block requests from datacenter IPs (Oracle Cloud, AWS, etc.) regardless of browser fingerprint or headed/headless mode. Best results come from running this on a residential connection.
  • Slow SPA rendering. For pages that progressively render content over many seconds, bump poll_budget_ms. The default returns the most-recent extraction at budget expiry, which on a still-rendering SPA may be partial.
  • Auth-walled content. This server uses a clean browser context with no cookies or stored auth. Logged-in pages won't work.

Comparison with related tools

  • @playwright/mcp (Microsoft) — general-purpose interactive browser automation: navigate, click, fill forms, run JS, take accessibility snapshots. Use it when you need to interact with a page. Use this server when you need to read a page.
  • mcp-server-fetch and similar HTTP-based servers — faster and lighter, but get blocked by Cloudflare and don't render JS. Try those first for compliant sites; reach for this one when they fail.
  • LM Studio Hub plugins like vadimfedenko/visit-website-reworked and npacker/web-tools — same idea inside LM Studio's plugin system. This server runs in any MCP client.

Roadmap

  • Persistent browser instance to amortize cold start
  • Accept: text/markdown fast path for Cloudflare's Markdown for Agents
  • Optional selector parameter to scope extraction
  • Optional viewport, user-agent, and locale overrides
  • Optional persistent browser context for sticky cookies / WAF trust
  • channel="chrome" option to use installed Google Chrome instead of bundled Chromium (further stealth for the hardest sites)

Contributing

Issues and PRs welcome. For substantive changes, please open an issue first to discuss the approach.

License

MIT — see LICENSE.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

web_to_markdown_mcp-0.5.0.tar.gz (10.8 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

web_to_markdown_mcp-0.5.0-py3-none-any.whl (9.7 kB view details)

Uploaded Python 3

File details

Details for the file web_to_markdown_mcp-0.5.0.tar.gz.

File metadata

  • Download URL: web_to_markdown_mcp-0.5.0.tar.gz
  • Upload date:
  • Size: 10.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for web_to_markdown_mcp-0.5.0.tar.gz
Algorithm Hash digest
SHA256 e46308e59fb6b137efd924df9b48a899378b093b305081ee5ec351765937d65c
MD5 0914ba54c69cfd63e5882d6e57f57b4d
BLAKE2b-256 4e859258827343104f2a91dd034ef6026eb49af1abd90dfd129903d7ce9e618b

See more details on using hashes here.

Provenance

The following attestation bundles were made for web_to_markdown_mcp-0.5.0.tar.gz:

Publisher: publish.yml on sidney/web-to-markdown-mcp

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file web_to_markdown_mcp-0.5.0-py3-none-any.whl.

File metadata

File hashes

Hashes for web_to_markdown_mcp-0.5.0-py3-none-any.whl
Algorithm Hash digest
SHA256 3fbd0c649e7d8bfadfa2ef6be598a1a3f1855296be0e746e3f58c37870ae6a89
MD5 b18ef92a1d43075b41b863b75583699e
BLAKE2b-256 f1ab6cfc55386432bb22cd19e555944aaa5282a50a7de22e8010d9a2bd26fa3b

See more details on using hashes here.

Provenance

The following attestation bundles were made for web_to_markdown_mcp-0.5.0-py3-none-any.whl:

Publisher: publish.yml on sidney/web-to-markdown-mcp

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page