Requirements
- Target platform
- OpenClaw
- Install method
- Manual import
- Extraction
- Extract archive
- Prerequisites
- OpenClaw
- Primary doc
- SKILL.md
Generate tech news digests with unified source model, quality scoring, and multi-format output. Five-layer data collection from RSS feeds, Twitter/X KOLs, Gi...
Generate tech news digests with unified source model, quality scoring, and multi-format output. Five-layer data collection from RSS feeds, Twitter/X KOLs, Gi...
Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.
I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Then review README.md for any prerequisites, environment setup, or post-install checks. Tell me what you changed and call out any manual steps you could not complete.
I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Then review README.md for any prerequisites, environment setup, or post-install checks. Summarize what changed and any follow-up checks I should run.
Automated tech news digest system with unified data source model, quality scoring pipeline, and template-based output generation.
Configuration Setup: Default configs are in config/defaults/. Copy to workspace for customization: mkdir -p workspace/config cp config/defaults/sources.json workspace/config/tech-news-digest-sources.json cp config/defaults/topics.json workspace/config/tech-news-digest-topics.json Environment Variables: TWITTERAPI_IO_KEY - twitterapi.io API key (optional, preferred) X_BEARER_TOKEN - Twitter/X official API bearer token (optional, fallback) TAVILY_API_KEY - Tavily Search API key, alternative to Brave (optional) WEB_SEARCH_BACKEND - Web search backend: auto|brave|tavily (optional, default: auto) BRAVE_API_KEYS - Brave Search API keys, comma-separated for rotation (optional) BRAVE_API_KEY - Single Brave key fallback (optional) GITHUB_TOKEN - GitHub personal access token (optional, improves rate limits) Generate Digest: # Unified pipeline (recommended) β runs all 6 sources in parallel + merge python3 scripts/run-pipeline.py \ --defaults config/defaults \ --config workspace/config \ --hours 48 --freshness pd \ --archive-dir workspace/archive/tech-news-digest/ \ --output /tmp/td-merged.json --verbose --force Use Templates: Apply Discord, email, or PDF templates to merged output
{ "sources": [ { "id": "openai-rss", "type": "rss", "name": "OpenAI Blog", "url": "https://openai.com/blog/rss.xml", "enabled": true, "priority": true, "topics": ["llm", "ai-agent"], "note": "Official OpenAI updates" }, { "id": "sama-twitter", "type": "twitter", "name": "Sam Altman", "handle": "sama", "enabled": true, "priority": true, "topics": ["llm", "frontier-tech"], "note": "OpenAI CEO" } ] }
{ "topics": [ { "id": "llm", "emoji": "π§ ", "label": "LLM / Large Models", "description": "Large Language Models, foundation models, breakthroughs", "search": { "queries": ["LLM latest news", "large language model breakthroughs"], "must_include": ["LLM", "large language model", "foundation model"], "exclude": ["tutorial", "beginner guide"] }, "display": { "max_items": 8, "style": "detailed" } } ] }
python3 scripts/run-pipeline.py \ --defaults config/defaults [--config CONFIG_DIR] \ --hours 48 --freshness pd \ --archive-dir workspace/archive/tech-news-digest/ \ --output /tmp/td-merged.json --verbose --force Features: Runs all 6 fetch steps in parallel, then merges + deduplicates + scores Output: Final merged JSON ready for report generation (~30s total) Metadata: Saves per-step timing and counts to *.meta.json GitHub Auth: Auto-generates GitHub App token if $GITHUB_TOKEN not set Fallback: If this fails, run individual scripts below
fetch-rss.py - RSS Feed Fetcher python3 scripts/fetch-rss.py [--defaults DIR] [--config DIR] [--hours 48] [--output FILE] [--verbose] Parallel fetching (10 workers), retry with backoff, feedparser + regex fallback Timeout: 30s per feed, ETag/Last-Modified caching fetch-twitter.py - Twitter/X KOL Monitor python3 scripts/fetch-twitter.py [--defaults DIR] [--config DIR] [--hours 48] [--output FILE] [--backend auto|official|twitterapiio] Backend auto-detection: uses twitterapi.io if TWITTERAPI_IO_KEY set, else official X API v2 if X_BEARER_TOKEN set Rate limit handling, engagement metrics, retry with backoff fetch-web.py - Web Search Engine python3 scripts/fetch-web.py [--defaults DIR] [--config DIR] [--freshness pd] [--output FILE] Auto-detects Brave API rate limit: paid plans β parallel queries, free β sequential Without API: generates search interface for agents fetch-github.py - GitHub Releases Monitor python3 scripts/fetch-github.py [--defaults DIR] [--config DIR] [--hours 168] [--output FILE] Parallel fetching (10 workers), 30s timeout Auth priority: $GITHUB_TOKEN β GitHub App auto-generate β gh CLI β unauthenticated (60 req/hr) fetch-github.py --trending - GitHub Trending Repos python3 scripts/fetch-github.py --trending [--hours 48] [--output FILE] [--verbose] Searches GitHub API for trending repos across 4 topics (LLM, AI Agent, Crypto, Frontier Tech) Quality scoring: base 5 + daily_stars_est / 10, max 15 fetch-reddit.py - Reddit Posts Fetcher python3 scripts/fetch-reddit.py [--defaults DIR] [--config DIR] [--hours 48] [--output FILE] Parallel fetching (4 workers), public JSON API (no auth required) 13 subreddits with score filtering enrich-articles.py - Article Full-Text Enrichment python3 scripts/enrich-articles.py --input merged.json --output enriched.json [--min-score 10] [--max-articles 15] [--verbose] Fetches full article text for high-scoring articles Cloudflare Markdown for Agents (preferred) β HTML extraction (fallback) β Skip (paywalled/social) Blog domain whitelist with lower score threshold (β₯3) Parallel fetching (5 workers, 10s timeout) merge-sources.py - Quality Scoring & Deduplication python3 scripts/merge-sources.py --rss FILE --twitter FILE --web FILE --github FILE --reddit FILE Quality scoring, title similarity dedup (85%), previous digest penalty Output: topic-grouped articles sorted by score validate-config.py - Configuration Validator python3 scripts/validate-config.py [--defaults DIR] [--config DIR] [--verbose] JSON schema validation, topic reference checks, duplicate ID detection generate-pdf.py - PDF Report Generator python3 scripts/generate-pdf.py --input report.md --output digest.pdf [--verbose] Converts markdown digest to styled A4 PDF with Chinese typography (Noto Sans CJK SC) Emoji icons, page headers/footers, blue accent theme. Requires weasyprint. sanitize-html.py - Safe HTML Email Converter python3 scripts/sanitize-html.py --input report.md --output email.html [--verbose] Converts markdown to XSS-safe HTML email with inline CSS URL whitelist (http/https only), HTML-escaped text content source-health.py - Source Health Monitor python3 scripts/source-health.py --rss FILE --twitter FILE --github FILE --reddit FILE --web FILE [--verbose] Tracks per-source success/failure history over 7 days Reports unhealthy sources (>50% failure rate) summarize-merged.py - Merged Data Summary python3 scripts/summarize-merged.py --input merged.json [--top N] [--topic TOPIC] Human-readable summary of merged data for LLM consumption Shows top articles per topic with scores and metrics
Place custom configs in workspace/config/ to override defaults: Sources: Append new sources, disable defaults with "enabled": false Topics: Override topic definitions, search queries, display settings Merge Logic: Sources with same id β user version takes precedence Sources with new id β appended to defaults Topics with same id β user version completely replaces default
// workspace/config/tech-news-digest-sources.json { "sources": [ { "id": "simonwillison-rss", "enabled": false, "note": "Disabled: too noisy for my use case" }, { "id": "my-custom-blog", "type": "rss", "name": "My Custom Tech Blog", "url": "https://myblog.com/rss", "enabled": true, "priority": true, "topics": ["frontier-tech"] } ] }
Bullet list format with link suppression (<link>) Mobile-optimized, emoji headers 2000 character limit awareness
Rich metadata, technical stats, archive links Executive summary, top articles section HTML-compatible formatting
A4 layout with Noto Sans CJK SC font for Chinese support Emoji icons, page headers/footers with page numbers Generated via scripts/generate-pdf.py (requires weasyprint)
RSS Feeds (62): AI labs, tech blogs, crypto news, Chinese tech media Twitter/X KOLs (48): AI researchers, crypto leaders, tech executives GitHub Repos (28): Major open-source projects (LangChain, vLLM, DeepSeek, Llama, etc.) Reddit (13): r/MachineLearning, r/LocalLLaMA, r/CryptoCurrency, r/ChatGPT, r/OpenAI, etc. Web Search (4 topics): LLM, AI Agent, Crypto, Frontier Tech All sources pre-configured with appropriate topic tags and priority levels.
pip install -r requirements.txt Optional but Recommended: feedparser>=6.0.0 - Better RSS parsing (fallback to regex if unavailable) jsonschema>=4.0.0 - Configuration validation All scripts work with Python 3.8+ standard library only.
# Validate configuration python3 scripts/validate-config.py --verbose # Test RSS feeds python3 scripts/fetch-rss.py --hours 1 --verbose # Check Twitter API python3 scripts/fetch-twitter.py --hours 1 --verbose
Digests automatically archived to <workspace>/archive/tech-news-digest/ Previous digest titles used for duplicate detection Old archives cleaned automatically (90+ days)
Network Failures: Retry with exponential backoff Rate Limits: Automatic retry with appropriate delays Invalid Content: Graceful degradation, detailed logging Configuration Errors: Schema validation with helpful messages
Set in ~/.zshenv or similar: # Twitter (at least one required for Twitter source) export TWITTERAPI_IO_KEY="your_key" # twitterapi.io key (preferred) export X_BEARER_TOKEN="your_bearer_token" # Official X API v2 (fallback) export TWITTER_API_BACKEND="auto" # auto|twitterapiio|official (default: auto) # Web Search (optional, enables web search layer) export WEB_SEARCH_BACKEND="auto" # auto|brave|tavily (default: auto) export TAVILY_API_KEY="tvly-xxx" # Tavily Search API (free 1000/mo) # Brave Search (alternative) export BRAVE_API_KEYS="key1,key2,key3" # Multiple keys, comma-separated rotation export BRAVE_API_KEY="key1" # Single key fallback export BRAVE_PLAN="free" # Override rate limit detection: free|pro # GitHub (optional, improves rate limits) export GITHUB_TOKEN="ghp_xxx" # PAT (simplest) export GH_APP_ID="12345" # Or use GitHub App for auto-token export GH_APP_INSTALL_ID="67890" export GH_APP_KEY_FILE="/path/to/key.pem" Twitter: TWITTERAPI_IO_KEY preferred ($3-5/mo); X_BEARER_TOKEN as fallback; auto mode tries twitterapiio first Web Search: Tavily (preferred in auto mode) or Brave; optional, fallback to agent web_search if unavailable GitHub: Auto-generates token from GitHub App if PAT not set; unauthenticated fallback (60 req/hr) Reddit: No API key needed (uses public JSON API)
This skill uses a prompt template pattern: the agent reads digest-prompt.md and follows its instructions. This is the standard OpenClaw skill execution model β the agent interprets structured instructions from skill-provided files. All instructions are shipped with the skill bundle and can be audited before installation.
The Python scripts make outbound requests to: RSS feed URLs (configured in tech-news-digest-sources.json) Twitter/X API (api.x.com or api.twitterapi.io) Brave Search API (api.search.brave.com) Tavily Search API (api.tavily.com) GitHub API (api.github.com) Reddit JSON API (reddit.com) No data is sent to any other endpoints. All API keys are read from environment variables declared in the skill metadata.
Email delivery uses send-email.py which constructs proper MIME multipart messages with HTML body + optional PDF attachment. Subject formats are hardcoded (Daily Tech Digest - YYYY-MM-DD). PDF generation uses generate-pdf.py via weasyprint. The prompt template explicitly prohibits interpolating untrusted content (article titles, tweet text, etc.) into shell arguments. Email addresses and subjects must be static placeholder values only.
Scripts read from config/ and write to workspace/archive/. No files outside the workspace are accessed.
RSS feeds failing: Check network connectivity, use --verbose for details Twitter rate limits: Reduce sources or increase interval Configuration errors: Run validate-config.py for specific issues No articles found: Check time window (--hours) and source enablement
All scripts support --verbose flag for detailed logging and troubleshooting.
Parallel Workers: Adjust MAX_WORKERS in scripts for your system Timeout Settings: Increase TIMEOUT for slow networks Article Limits: Adjust MAX_ARTICLES_PER_FEED based on needs
The digest prompt instructs agents to run Python scripts via shell commands. All script paths and arguments are skill-defined constants β no user input is interpolated into commands. Two scripts use subprocess: run-pipeline.py orchestrates child fetch scripts (all within scripts/ directory) fetch-github.py has two subprocess calls: openssl dgst -sha256 -sign for JWT signing (only if GH_APP_* env vars are set β signs a self-constructed JWT payload, no user content involved) gh auth token CLI fallback (only if gh is installed β reads from gh's own credential store) No user-supplied or fetched content is ever interpolated into subprocess arguments. Email delivery uses send-email.py which builds MIME messages programmatically β no shell interpolation. PDF generation uses generate-pdf.py via weasyprint. Email subjects are static format strings only β never constructed from fetched data.
Scripts do not directly read ~/.config/, ~/.ssh/, or any credential files. All API tokens are read from environment variables declared in the skill metadata. The GitHub auth cascade is: $GITHUB_TOKEN env var (you control what to provide) GitHub App token generation (only if you set GH_APP_ID, GH_APP_INSTALL_ID, and GH_APP_KEY_FILE β uses inline JWT signing via openssl CLI, no external scripts involved) gh auth token CLI (delegates to gh's own secure credential store) Unauthenticated (60 req/hr, safe fallback) If you prefer no automatic credential discovery, simply set $GITHUB_TOKEN and the script will use it directly without attempting steps 2-3.
This skill does not install any packages. requirements.txt lists optional dependencies (feedparser, jsonschema) for reference only. All scripts work with Python 3.8+ standard library. Users should install optional deps in a virtualenv if desired β the skill never runs pip install.
URL resolution rejects non-HTTP(S) schemes (javascript:, data:, etc.) RSS fallback parsing uses simple, non-backtracking regex patterns (no ReDoS risk) All fetched content is treated as untrusted data for display only
Scripts make outbound HTTP requests to configured RSS feeds, Twitter API, GitHub API, Reddit JSON API, Brave Search API, and Tavily Search API. No inbound connections or listeners are created.
Agent frameworks, memory systems, reasoning layers, and model-native orchestration.
Largest current source with strong distribution and engagement signals.