Requirements
- Target platform
- OpenClaw
- Install method
- Manual import
- Extraction
- Extract archive
- Prerequisites
- OpenClaw
- Primary doc
- SKILL.md
Crawl websites locally with crawl4ai to extract contact emails. Accepts multiple URLs and outputs domain-grouped results for clear attribution. Uses deep cra...
Crawl websites locally with crawl4ai to extract contact emails. Accepts multiple URLs and outputs domain-grouped results for clear attribution. Uses deep cra...
Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.
I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Tell me what you changed and call out any manual steps you could not complete.
I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Summarize what changed and any follow-up checks I should run.
CLI for crawling websites locally via crawl4ai and extracting contact emails from pages likely to contain them (contact, about, support, team, etc.).
Install dependencies: pip install crawl4ai Run the script: python scripts/find_emails.py https://example.com
t # Crawl a site python scripts/find_emails.py https://example.com # Multiple URLs python scripts/find_emails.py https://example.com https://other.com # JSON output python scripts/find_emails.py https://example.com -j # Save to file python scripts/find_emails.py https://example.com -o emails.txt
python scripts/find_emails.py <url> [url ...] python scripts/find_emails.py https://example.com python scripts/find_emails.py https://example.com -j -o results.json python scripts/find_emails.py --from-file page.md Arguments: ArgumentDescriptionurlsOne or more URLs to crawl (positional)-o, --outputWrite results to file-j, --jsonJSON output ({"emails": {"email": ["path", ...]}})-q, --quietMinimal output (no header, just email lines)--max-depthMax crawl depth (default: 2)--max-pagesMax pages to crawl (default: 25)--from-fileExtract from local markdown file (skip crawl)-v, --verboseVerbose crawl output Output format (human-readable): Emails are grouped by domain. Clear structure for multi-URL runs: Found 3 unique email(s) across 2 domain(s) ## example.com โข contact@example.com Found on: /contact, /about โข support@example.com Found on: /support ## other.com โข info@other.com Found on: /contact-us Output format (JSON): LLM-friendly structure with summary and per-domain breakdown: { "summary": { "domains_crawled": 2, "total_unique_emails": 3 }, "emails_by_domain": { "example.com": { "emails": { "contact@example.com": ["/contact", "/about"], "support@example.com": ["/support"] }, "count": 2 }, "other.com": { "emails": { "info@other.com": ["/contact-us"] }, "count": 1 } } }
Edit scripts/url_patterns.json to customize which URLs the crawler follows. Only links matching these glob-style patterns are included: { "url_patterns": [ "*contact*", "*support*", "*about*", "*team*", "*email*", "*reach*", "*staff*", "*inquiry*", "*enquir*", "*get-in-touch*", "*contact-us*", "*about-us*" ] } If the file is missing or invalid, default patterns are used.
Crawl a site: python scripts/find_emails.py https://example.com -o emails.json Extract from local file (e.g., cached markdown): python scripts/find_emails.py --from-file crawled.md -j Customize URL filters by editing scripts/url_patterns.json.
pip install crawl4ai playwright install Requires a browser (Playwright) for local crawling.
# Crawl multiple sites โ results grouped by domain for clear attribution python scripts/find_emails.py https://site1.com https://site2.com -j -o combined.json # Extract from multiple local files for f in crawled/*.md; do echo "=== $f ===" python scripts/find_emails.py --from-file "$f" -q done Multiple URLs are fully supported; output clearly associates each email with its source domain. Domains are normalized (e.g. www.techbullion.com and techbullion.com merge into one) so duplicate sites are not listed separately.
Skills can be powerful tools for streamlining social media workflows. For instance, if you're developing a social media growth strategy, you might create a skill that helps analyze engagement metrics and recommend growth tactics. Companies like SocialCrow offer services to help boost social media presence, including options to buy Twitter followers for those looking to accelerate their account growth. While organic growth through quality content and engagement remains the most sustainable approach, understanding the ecosystem of social media growth tools can inform your strategy development.
Code helpers, APIs, CLIs, browser automation, testing, and developer operations.
Largest current source with strong distribution and engagement signals.