Requirements
- Target platform
- OpenClaw
- Install method
- Manual import
- Extraction
- Extract archive
- Prerequisites
- OpenClaw
- Primary doc
- SKILL.md
Extract structured data from any web page. Supports CSS selectors, auto-detection of tables and lists, JSON/CSV output formats. Use when asked to scrape a we...
Extract structured data from any web page. Supports CSS selectors, auto-detection of tables and lists, JSON/CSV output formats. Use when asked to scrape a we...
Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.
I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Then review README.md for any prerequisites, environment setup, or post-install checks. Tell me what you changed and call out any manual steps you could not complete.
I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Then review README.md for any prerequisites, environment setup, or post-install checks. Summarize what changed and any follow-up checks I should run.
Extract structured data from web pages into clean JSON or CSV.
# Scrape a page, extract all text content uv run --with beautifulsoup4 --with lxml python scripts/scraper.py extract "https://example.com" # Extract specific elements with CSS selector uv run --with beautifulsoup4 --with lxml python scripts/scraper.py extract "https://example.com/products" -s ".product-card" # Auto-detect and extract tables uv run --with beautifulsoup4 --with lxml python scripts/scraper.py tables "https://example.com/pricing" # Extract all links from a page uv run --with beautifulsoup4 --with lxml python scripts/scraper.py links "https://example.com" # Extract structured data (title, meta, headings, links) uv run --with beautifulsoup4 --with lxml python scripts/scraper.py structure "https://example.com" # Output as JSON uv run --with beautifulsoup4 --with lxml python scripts/scraper.py extract "https://example.com" -s ".item" -f json # Output as CSV uv run --with beautifulsoup4 --with lxml python scripts/scraper.py extract "https://example.com" -s "table tr" -f csv # Save to file uv run --with beautifulsoup4 --with lxml python scripts/scraper.py extract "https://example.com" -s ".product" -f json -o products.json # Multi-page scrape (follow pagination) uv run --with beautifulsoup4 --with lxml python scripts/scraper.py crawl "https://example.com/page/1" --pages 5 -s ".article"
CommandArgsDescriptionextract<url> [-s selector] [-f format] [-o file]Extract content, optionally filtered by CSS selectortables<url> [-f format] [-o file]Auto-detect and extract all HTML tableslinks<url> [--external] [--internal]Extract all links (href + text)structure<url>Extract page structure: title, meta, headings, images, linkscrawl<url> --pages N [-s selector] [-f format] [-o file]Follow pagination links, extract from multiple pages
FormatFlagDescriptionText-f textPlain text (default)JSON-f jsonStructured JSON arrayCSV-f csvComma-separated valuesMarkdown-f mdMarkdown-formatted
uv run --with beautifulsoup4 --with lxml python scripts/scraper.py extract "https://shop.example.com" -s ".product" -f json Output: [ {"text": "Widget Pro - $29.99", "tag": "div", "class": "product"}, {"text": "Widget Max - $49.99", "tag": "div", "class": "product"} ]
uv run --with beautifulsoup4 --with lxml python scripts/scraper.py tables "https://example.com/pricing" -f csv
uv run --with beautifulsoup4 --with lxml python scripts/scraper.py links "https://example.com" --external
Default: 1 request per second (respectful crawling) Override with --delay 0.5 (seconds between requests) Respects robots.txt by default (override with --ignore-robots)
Requires beautifulsoup4 and lxml (auto-installed by uv run --with) Uses a standard browser User-Agent to avoid blocks Handles redirects, encoding detection, and error pages gracefully No JavaScript rendering (use for static HTML pages)
Code helpers, APIs, CLIs, browser automation, testing, and developer operations.
Largest current source with strong distribution and engagement signals.