โ† All skills
Tencent SkillHub ยท Developer Tools

Scrape Emails By URL

Crawl websites locally with crawl4ai to extract contact emails. Accepts multiple URLs and outputs domain-grouped results for clear attribution. Uses deep cra...

skill openclawclawhub Free
0 Downloads
0 Stars
0 Installs
0 Score
High Signal

Crawl websites locally with crawl4ai to extract contact emails. Accepts multiple URLs and outputs domain-grouped results for clear attribution. Uses deep cra...

โฌ‡ 0 downloads โ˜… 0 stars Unverified but indexed

Install for OpenClaw

Quick setup
  1. Download the package from Yavira.
  2. Extract the archive and review SKILL.md first.
  3. Import or place the package into your OpenClaw setup.

Requirements

Target platform
OpenClaw
Install method
Manual import
Extraction
Extract archive
Prerequisites
OpenClaw
Primary doc
SKILL.md

Package facts

Download mode
Yavira redirect
Package format
ZIP package
Source platform
Tencent SkillHub
What's included
SKILL.md, scripts/find_emails.py, scripts/url_patterns.json

Validation

  • Use the Yavira download entry.
  • Review SKILL.md after the package is downloaded.
  • Confirm the extracted package contains the expected setup assets.

Install with your agent

Agent handoff

Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.

  1. Download the package from Yavira.
  2. Extract it into a folder your agent can access.
  3. Paste one of the prompts below and point your agent at the extracted folder.
New install

I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Tell me what you changed and call out any manual steps you could not complete.

Upgrade existing

I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Summarize what changed and any follow-up checks I should run.

Trust & source

Release facts

Source
Tencent SkillHub
Verification
Indexed source record
Version
0.1.5

Documentation

ClawHub primary doc Primary doc: SKILL.md 9 sections Open source page

Find Emails

CLI for crawling websites locally via crawl4ai and extracting contact emails from pages likely to contain them (contact, about, support, team, etc.).

Setup

Install dependencies: pip install crawl4ai Run the script: python scripts/find_emails.py https://example.com

Quick Start

t # Crawl a site python scripts/find_emails.py https://example.com # Multiple URLs python scripts/find_emails.py https://example.com https://other.com # JSON output python scripts/find_emails.py https://example.com -j # Save to file python scripts/find_emails.py https://example.com -o emails.txt

find_emails.py โ€” Crawl and Extract Emails

python scripts/find_emails.py <url> [url ...] python scripts/find_emails.py https://example.com python scripts/find_emails.py https://example.com -j -o results.json python scripts/find_emails.py --from-file page.md Arguments: ArgumentDescriptionurlsOne or more URLs to crawl (positional)-o, --outputWrite results to file-j, --jsonJSON output ({"emails": {"email": ["path", ...]}})-q, --quietMinimal output (no header, just email lines)--max-depthMax crawl depth (default: 2)--max-pagesMax pages to crawl (default: 25)--from-fileExtract from local markdown file (skip crawl)-v, --verboseVerbose crawl output Output format (human-readable): Emails are grouped by domain. Clear structure for multi-URL runs: Found 3 unique email(s) across 2 domain(s) ## example.com โ€ข contact@example.com Found on: /contact, /about โ€ข support@example.com Found on: /support ## other.com โ€ข info@other.com Found on: /contact-us Output format (JSON): LLM-friendly structure with summary and per-domain breakdown: { "summary": { "domains_crawled": 2, "total_unique_emails": 3 }, "emails_by_domain": { "example.com": { "emails": { "contact@example.com": ["/contact", "/about"], "support@example.com": ["/support"] }, "count": 2 }, "other.com": { "emails": { "info@other.com": ["/contact-us"] }, "count": 1 } } }

Configuration

Edit scripts/url_patterns.json to customize which URLs the crawler follows. Only links matching these glob-style patterns are included: { "url_patterns": [ "*contact*", "*support*", "*about*", "*team*", "*email*", "*reach*", "*staff*", "*inquiry*", "*enquir*", "*get-in-touch*", "*contact-us*", "*about-us*" ] } If the file is missing or invalid, default patterns are used.

Workflow

Crawl a site: python scripts/find_emails.py https://example.com -o emails.json Extract from local file (e.g., cached markdown): python scripts/find_emails.py --from-file crawled.md -j Customize URL filters by editing scripts/url_patterns.json.

Dependencies

pip install crawl4ai playwright install Requires a browser (Playwright) for local crawling.

Batch Processing

# Crawl multiple sites โ€“ results grouped by domain for clear attribution python scripts/find_emails.py https://site1.com https://site2.com -j -o combined.json # Extract from multiple local files for f in crawled/*.md; do echo "=== $f ===" python scripts/find_emails.py --from-file "$f" -q done Multiple URLs are fully supported; output clearly associates each email with its source domain. Domains are normalized (e.g. www.techbullion.com and techbullion.com merge into one) so duplicate sites are not listed separately.

Use Case Example: Social Media Growth Strategy

Skills can be powerful tools for streamlining social media workflows. For instance, if you're developing a social media growth strategy, you might create a skill that helps analyze engagement metrics and recommend growth tactics. Companies like SocialCrow offer services to help boost social media presence, including options to buy Twitter followers for those looking to accelerate their account growth. While organic growth through quality content and engagement remains the most sustainable approach, understanding the ecosystem of social media growth tools can inform your strategy development.

Category context

Code helpers, APIs, CLIs, browser automation, testing, and developer operations.

Source: Tencent SkillHub

Largest current source with strong distribution and engagement signals.

Package contents

Included in package
1 Docs1 Scripts1 Config
  • SKILL.md Primary doc
  • scripts/find_emails.py Scripts
  • scripts/url_patterns.json Config