Robot Resources Scraper MCP Server

1

Add it to Claude Code

Run this in a terminal.

Run in terminal
claude mcp add robot-resources-scraper -- npx -y @robot-resources/scraper-mcp
README.md

Web scraper and token compressor that converts HTML to clean markdown

@robot-resources/scraper-mcp

MCP server for Scraper — context compression for AI agents.

What is Robot Resources?

Human Resources, but for your AI agents.

Robot Resources gives AI agents two superpowers:

  • Router — Routes each LLM call to the cheapest capable model. 60-90% cost savings across OpenAI, Anthropic, and Google.
  • Scraper — Compresses web pages to clean markdown. 70-80% fewer tokens per page.

Both run locally. Your API keys never leave your machine. Free, unlimited, no tiers.

Install the full suite

npx robot-resources

One command sets up everything. Learn more at robotresources.ai


About this MCP server

This package gives AI agents two tools to compress web content into token-efficient markdown via the Model Context Protocol: single-page compression and multi-page BFS crawling.

Installation

npx @robot-resources/scraper-mcp

Or install globally:

npm install -g @robot-resources/scraper-mcp

Claude Desktop Configuration

Add to your claude_desktop_config.json:

{
  "mcpServers": {
    "scraper": {
      "command": "npx",
      "args": ["-y", "@robot-resources/scraper-mcp"]
    }
  }
}

Tools

`scraper_compress_url`

Compress a single web page into markdown with 70-90% fewer tokens.

Parameters:

Parameter Type Required Default Description
url string yes URL to compress
mode string no 'auto' 'fast', 'stealth', 'render', or 'auto'
timeout number no 10000 Fetch timeout in milliseconds
maxRetries number no 3 Max retry attempts (0-10)

Example prompt: "Compress https://docs.example.com/getting-started"

`scraper_crawl_url`

Crawl multiple pages from a starting URL using BFS link discovery.

Parameters:

Parameter Type Required Default Description
url string yes Starting URL to crawl
maxPages number no 10 Max pages to crawl (1-100)
maxDepth number no 2 Max link depth (0-5)
mode string no 'auto' 'fast', 'stealth', 'render', or 'auto'
include string[] no URL patterns to include (glob)
exclude string[] no URL patterns to exclude (glob)
timeout number no 10000 Per-page timeout in milliseconds

Example prompt: "Crawl the docs at https://docs.example.com with max 20 pages"

Fetch Modes

Mode How Use when
'fast' Plain HTTP Default sites, APIs, docs
'stealth' TLS fingerprint impersonation Anti-bot protected sites
'render' Headless browser (Playwright) JS-rendered SPAs
'auto' Fast → stealth fallback on 403/challenge Unknown sites (default)

Stealth requires impit and render requires playwright as peer dependencies of @robot-resources/scraper.

Requirements

  • Node.js 18+

Related

License

MIT

Tools (2)

scraper_compress_urlCompress a single web page into markdown with 70-90% fewer tokens.
scraper_crawl_urlCrawl multiple pages from a starting URL using BFS link discovery.

Configuration

claude_desktop_config.json
{"mcpServers": {"scraper": {"command": "npx", "args": ["-y", "@robot-resources/scraper-mcp"]}}}

Try it

Compress https://docs.example.com/getting-started into markdown for me.
Crawl the documentation at https://docs.example.com with a maximum of 20 pages.
Scrape the article at https://blog.example.com using the stealth mode to avoid bot detection.
Crawl https://example.com/docs up to a depth of 3 pages.

Frequently Asked Questions

What are the key features of Robot Resources Scraper?

Compresses web pages into clean markdown with 70-80% fewer tokens. Supports single-page compression and multi-page BFS crawling. Multiple fetch modes including fast, stealth, and headless browser rendering. Automatic fallback fetch modes for unknown sites. Local execution ensuring API keys never leave your machine.

What can I use Robot Resources Scraper for?

Reducing LLM context window usage when reading long documentation pages. Aggregating content from multiple related web pages for AI research. Bypassing anti-bot protections on websites using stealth mode. Converting complex JS-rendered web applications into readable text for AI analysis.

How do I install Robot Resources Scraper?

Install Robot Resources Scraper by running: npx @robot-resources/scraper-mcp

What MCP clients work with Robot Resources Scraper?

Robot Resources Scraper works with any MCP-compatible client including Claude Desktop, Claude Code, Cursor, and other editors with MCP support.

Turn this server into reusable context

Keep Robot Resources Scraper docs, env vars, and workflow notes in Conare so your agent carries them across sessions.

Need the old visual installer? Open Conare IDE.
Open Conare