Requirements
- Target platform
- OpenClaw
- Install method
- Manual import
- Extraction
- Extract archive
- Prerequisites
- OpenClaw
- Primary doc
- SKILL.md
URL content extraction via Parallel API. Extracts clean markdown from webpages, articles, PDFs, and JS-heavy sites. Use for reading specific URLs with LLM-ready output.
URL content extraction via Parallel API. Extracts clean markdown from webpages, articles, PDFs, and JS-heavy sites. Use for reading specific URLs with LLM-ready output.
Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.
I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Tell me what you changed and call out any manual steps you could not complete.
I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Summarize what changed and any follow-up checks I should run.
Extract clean, LLM-ready content from URLs. Handles webpages, articles, PDFs, and JavaScript-heavy sites that need rendering.
Trigger this skill when the user asks for: "read this URL", "fetch this page", "extract from..." "get the content from [URL]" "what does this article say?" Reading PDFs, JS-heavy pages, or paywalled content Getting clean markdown from messy web pages Use Search to discover; use Extract to read.
parallel-cli extract "https://example.com/article" --json
parallel-cli extract "<url>" [options]
FlagDescription--url "<url>"URL to extract (repeatable, max 10)--objective "<focus>"Focus extraction on specific content--jsonOutput as JSON--excerpts / --no-excerptsInclude relevant excerpts (default: on)--full-content / --no-full-contentInclude full page content--excerpts-max-chars NMax chars per excerpt--excerpts-max-total-chars NMax total excerpt chars--full-max-chars NMax full content chars-o <file>Save output to file
Basic extraction: parallel-cli extract "https://example.com/article" --json Focused extraction: parallel-cli extract "https://example.com/pricing" \ --objective "pricing tiers and features" \ --json Full content for PDFs: parallel-cli extract "https://example.com/whitepaper.pdf" \ --full-content \ --json Multiple URLs: parallel-cli extract \ --url "https://example.com/page1" \ --url "https://example.com/page2" \ --json
Search with an objective + keyword queries Inspect titles/URLs/dates; choose the best sources Extract the specific pages you need (top N URLs) Answer using the extracted excerpts/content
When extracting, provide context: What specific information you're looking for Why you need it (helps focus extraction) Good: --objective "Find the installation steps and system requirements" Poor: --objective "Read the page"
Returns structured JSON with: url โ source URL title โ page title excerpts[] โ relevant text excerpts (if enabled) full_content โ complete page content (if enabled) publish_date โ when available
When turning extracted content into a user-facing answer: Keep content verbatim โ do not paraphrase unnecessarily Extract ALL list items exhaustively Strip noise: nav menus, footers, ads, "click here" links Preserve all facts, names, numbers, dates, quotes Include URL + publish_date for transparency
For long conversations, save results and use sessions_spawn: parallel-cli extract "<url>" --json -o /tmp/extract-<topic>.json Then spawn a sub-agent: { "tool": "sessions_spawn", "task": "Read /tmp/extract-<topic>.json and summarize the key content.", "label": "extract-summary" }
Exit CodeMeaning0Success1Unexpected error (network, parse)2Invalid arguments3API error (non-2xx)
Get an API key at parallel.ai Install the CLI: curl -fsSL https://parallel.ai/install.sh | bash export PARALLEL_API_KEY=your-key
API Docs Extract API Reference
Code helpers, APIs, CLIs, browser automation, testing, and developer operations.
Largest current source with strong distribution and engagement signals.