{
  "schemaVersion": "1.0",
  "item": {
    "slug": "solo-research",
    "name": "Research",
    "source": "tencent",
    "type": "skill",
    "category": "开发工具",
    "sourceUrl": "https://clawhub.ai/fortunto2/solo-research",
    "canonicalUrl": "https://clawhub.ai/fortunto2/solo-research",
    "targetPlatform": "OpenClaw"
  },
  "install": {
    "downloadMode": "redirect",
    "downloadUrl": "/downloads/solo-research",
    "sourceDownloadUrl": "https://wry-manatee-359.convex.site/api/v1/download?slug=solo-research",
    "sourcePlatform": "tencent",
    "targetPlatform": "OpenClaw",
    "installMethod": "Manual import",
    "extraction": "Extract archive",
    "prerequisites": [
      "OpenClaw"
    ],
    "packageFormat": "ZIP package",
    "includedAssets": [
      "SKILL.md",
      "references/domain-check.md",
      "references/research-template.md"
    ],
    "primaryDoc": "SKILL.md",
    "quickSetup": [
      "Download the package from Yavira.",
      "Extract the archive and review SKILL.md first.",
      "Import or place the package into your OpenClaw setup."
    ],
    "agentAssist": {
      "summary": "Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.",
      "steps": [
        "Download the package from Yavira.",
        "Extract it into a folder your agent can access.",
        "Paste one of the prompts below and point your agent at the extracted folder."
      ],
      "prompts": [
        {
          "label": "New install",
          "body": "I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Tell me what you changed and call out any manual steps you could not complete."
        },
        {
          "label": "Upgrade existing",
          "body": "I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Summarize what changed and any follow-up checks I should run."
        }
      ]
    },
    "sourceHealth": {
      "source": "tencent",
      "slug": "solo-research",
      "status": "healthy",
      "reason": "direct_download_ok",
      "recommendedAction": "download",
      "checkedAt": "2026-05-08T05:02:08.449Z",
      "expiresAt": "2026-05-15T05:02:08.449Z",
      "httpStatus": 200,
      "finalUrl": "https://wry-manatee-359.convex.site/api/v1/download?slug=solo-research",
      "contentType": "application/zip",
      "probeMethod": "head",
      "details": {
        "probeUrl": "https://wry-manatee-359.convex.site/api/v1/download?slug=solo-research",
        "contentDisposition": "attachment; filename=\"solo-research-1.7.1.zip\"",
        "redirectLocation": null,
        "bodySnippet": null,
        "slug": "solo-research"
      },
      "scope": "item",
      "summary": "Item download looks usable.",
      "detail": "Yavira can redirect you to the upstream package for this item.",
      "primaryActionLabel": "Download for OpenClaw",
      "primaryActionHref": "/downloads/solo-research"
    },
    "validation": {
      "installChecklist": [
        "Use the Yavira download entry.",
        "Review SKILL.md after the package is downloaded.",
        "Confirm the extracted package contains the expected setup assets."
      ],
      "postInstallChecks": [
        "Confirm the extracted package includes the expected docs or setup files.",
        "Validate the skill or prompts are available in your target agent workspace.",
        "Capture any manual follow-up steps the agent could not complete."
      ]
    },
    "downloadPageUrl": "https://openagent3.xyz/downloads/solo-research",
    "agentPageUrl": "https://openagent3.xyz/skills/solo-research/agent",
    "manifestUrl": "https://openagent3.xyz/skills/solo-research/agent.json",
    "briefUrl": "https://openagent3.xyz/skills/solo-research/agent.md"
  },
  "agentAssist": {
    "summary": "Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.",
    "steps": [
      "Download the package from Yavira.",
      "Extract it into a folder your agent can access.",
      "Paste one of the prompts below and point your agent at the extracted folder."
    ],
    "prompts": [
      {
        "label": "New install",
        "body": "I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Tell me what you changed and call out any manual steps you could not complete."
      },
      {
        "label": "Upgrade existing",
        "body": "I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Summarize what changed and any follow-up checks I should run."
      }
    ]
  },
  "documentation": {
    "source": "clawhub",
    "primaryDoc": "SKILL.md",
    "sections": [
      {
        "title": "/research",
        "body": "Deep research before PRD generation. Produces a structured research.md with competitive analysis, user pain points, SEO/ASO keywords, naming/domain options, and market sizing."
      },
      {
        "title": "MCP Tools (use if available)",
        "body": "If MCP tools are available, prefer them over CLI:\n\nkb_search(query, n_results) — search knowledge base for related docs\nweb_search(query, engines, include_raw_content) — web search with engine routing\nsession_search(query, project) — find how similar research was done before\nproject_info(name) — check project details and stacks\ncodegraph_explain(project) — architecture overview of an existing project (stack, patterns, deps)\ncodegraph_query(query) — raw Cypher queries against code graph (find shared packages, dependencies)\nproject_code_search(query, project) — semantic search over project source code\n\nMCP web_search supports engine override: engines=\"reddit\", engines=\"youtube\", etc.\nIf MCP tools are not available, use WebSearch/WebFetch as primary. If MCP web_search tool is available, use it for better results."
      },
      {
        "title": "Reddit Search Best Practices",
        "body": "Max 3 keywords in reddit queries — more keywords = fewer results\nGood: \"product hunt outreach launch\" — Bad: \"product hunt scraper maker profiles linkedin outreach launch strategy\"\ninclude_raw_content=true rarely works for Reddit — use fallback chain below"
      },
      {
        "title": "Reddit Content Access — Fallback Chain",
        "body": "When a search finds a relevant Reddit post, reading its full content requires a fallback chain:\n\n1. MCP Playwright (old.reddit.com)     ← BEST: bypasses CAPTCHA, full post + comments\n2. PullPush API (api.pullpush.io)      ← search by query/subreddit/author/score/date\n3. MCP web_search include_raw_content   ← sometimes works, often truncated\n4. WebFetch / WebSearch snippets        ← last resort, partial data only\n\nMethod 1: MCP Playwright (recommended for full post content)\n\nUse browser_navigate(\"https://old.reddit.com/r/...\") — old.reddit.com loads without CAPTCHA\nwww.reddit.com shows CAPTCHA (\"Prove your humanity\"), always use old.reddit.com\nSnapshot contains full post text + comments in structured YAML\nExample: old.reddit.com/r/indiehackers/comments/abc123/post_title/\n\nMethod 2: PullPush API (for search/discovery)\n\nEndpoint: https://api.pullpush.io/reddit/submission/search\nParams: q, subreddit, author, score (e.g. >10,<100), since/until (unix timestamps), size (max 100)\nRate limits: soft 15 req/min, hard 30 req/min, 1000 req/hr. Sleep 4 sec between requests.\nReturns JSON with full selftext, author, score, created_utc\nComment search: /reddit/comment/search (same params)\nCan use via curl:\n\ncurl -s \"https://api.pullpush.io/reddit/submission/search?q=product+hunt+launch&subreddit=indiehackers&size=10\"\n\nMethod 3: Reddit .json endpoint (often blocked)\n\nAppend .json to any Reddit URL: reddit.com/r/sub/comments/id.json\nReturns raw JSON with full post + comments\nFrequently blocked (403/429) — use as opportunistic fallback only\n\nMethod 4: PRAW (Reddit Official API, for live search/user profiles)\n\npraw-dev/praw — Python Reddit API Wrapper\nOAuth2 auth, built-in rate limiting, sync/async support\nBest for: live subreddit search, user profiles, comment trees\npip install praw / uv add praw"
      },
      {
        "title": "Search Strategy: Hybrid (MCP + WebSearch)",
        "body": "Use multiple search backends together. Each has strengths:\n\nStepBest backendWhyCompetitorsWebSearch + site:producthunt.com + site:g2.comBroad discovery + Product Hunt + B2B reviewsReddit / Pain pointsMCP web_search with engines: reddit (max 3 keywords!) + MCP Playwright for full postsPullPush API, selftext in contentYouTube reviewsMCP web_search with engines: youtubeVideo reviews (views = demand)Market sizeWebSearchSynthesizes numbers from 10 sourcesSEO / ASOWebSearchBroader coverage, trend dataPage scrapingWebFetch or MCP web_search with include_raw_contentUp to 5000 chars of page contentHacker NewsWebSearch site:news.ycombinator.comHN discussions and opinionsFunding / CompaniesWebSearch site:crunchbase.comCompetitor funding, team sizeVerified revenueWebFetch trustmrr.com/startup/<slug>Stripe-verified MRR, growth, tech stack, traffic"
      },
      {
        "title": "Search Availability",
        "body": "Use WebSearch/WebFetch as primary. If MCP web_search tool is available, use it for better results (supports engine routing and raw content extraction)."
      },
      {
        "title": "Steps",
        "body": "Parse the idea from $ARGUMENTS. If empty, ask the user what idea they want to research.\n\n\nDetect product type — infer from the idea description:\n\nKeywords like \"app\", \"mobile\", \"iPhone\", \"Android\" → mobile (ios/android)\nKeywords like \"website\", \"SaaS\", \"dashboard\", \"web app\" → web\nKeywords like \"CLI\", \"terminal\", \"command line\" → cli\nKeywords like \"API\", \"backend\", \"service\" → api\nKeywords like \"extension\", \"plugin\", \"browser\" → web (extension)\nDefault if unclear → web\nOnly ask via AskUserQuestion if truly ambiguous (e.g., \"build a todo app\" could be web or mobile)\nThis determines which research sections apply (ASO for mobile, SEO for web, etc.)\n\n\n\nSearch knowledge base and past work:\n\nIf MCP kb_search available: kb_search(query=\"<idea keywords>\", n_results=5)\nIf MCP session_search available: session_search(query=\"<idea keywords>\") — check if this idea was researched before\nOtherwise: Grep for keywords in .md files\nCheck if research.md or prd.md already exist for this idea.\n\n\n\nCheck existing portfolio (if MCP codegraph tools available):\n\ncodegraph_explain(project=\"<similar project>\") — architecture overview of related projects in the portfolio\nproject_code_search(query=\"<relevant pattern>\", project=\"<sibling>\") — find reusable code, patterns, infrastructure\ncodegraph_query(\"MATCH (p:Project)-[:DEPENDS_ON]->(pkg:Package) WHERE pkg.name CONTAINS '<relevant tech>' RETURN p.name, pkg.name\") — find projects using similar tech\nThis helps assess: feasibility, reusable code, stack decisions, and time estimates\nIf no MCP tools available, skip this step.\n\n\n\nCompetitive analysis — use WebSearch (primary) + MCP web_search (if available):\n\n\"<idea> competitors alternatives 2026\" — broad discovery\n\"<idea> app review pricing\" — pricing data\nWebFetch or MCP include_raw_content=true: scrape competitor URLs for detailed pricing\nMCP engines: reddit or WebSearch: \"<idea> vs\" — user opinions\n\"site:producthunt.com <idea>\" — Product Hunt launches\n\"site:g2.com <idea>\" or \"site:capterra.com <idea>\" — B2B reviews\n\"site:crunchbase.com <competitor>\" — funding, team size\n\"site:trustmrr.com <idea>\" or WebFetch trustmrr.com/startup/<slug> — Stripe-verified MRR, growth %, tech stack, traffic (24h/7d/30d)\nFor each competitor extract: name, URL, pricing, key features, weaknesses, verified MRR (if on TrustMRR)\n\n\n\nUser pain points — use MCP web_search / WebSearch + YouTube:\n\nMCP engines: reddit or WebSearch: \"<problem>\" — Reddit discussions (max 3 keywords!)\nIf Reddit post found but content not available → open via MCP Playwright: browser_navigate(\"https://old.reddit.com/r/...\") — old.reddit.com bypasses CAPTCHA\nMCP engines: youtube or WebSearch: \"<problem> review\" — video reviews\n\"site:news.ycombinator.com <problem>\" — Hacker News opinions\nWebSearch: \"<problem> frustrating OR annoying\" — broader sweep\nSynthesis: top 5 pain points with quotes and source URLs\n\n\n\nSEO / ASO analysis (depends on product type from step 2):\nFor web apps:\n\n\"<competitor> SEO keywords ranking\" — competitor keywords\n\"<problem domain> search volume trends 2026\" — demand signals\nWebFetch or MCP include_raw_content: scrape competitor pages for meta tags\nResult: keyword table (keyword, intent, competition, relevance)\n\nFor mobile apps:\n\n\"<category> App Store top apps keywords 2026\" — category landscape\n\"site:reddit.com <competitor app> review\" — user complaints\nResult: ASO keywords, competitor ratings, common complaints\n\n\n\nNaming, domains, and company registration:\n\nGenerate 7-10 name candidates (mix of descriptive + invented/brandable)\nDomain availability: triple verification (whois → dig → RDAP)\nTrademark + company name conflict checks\n\nSee references/domain-check.md (bundled with this skill) for TLD priority tiers, bash scripts, gotchas, and trademark check methods.\n\n\nMarket sizing (TAM/SAM/SOM) — use WebSearch (primary):\n\nWebSearch: \"<market> market size 2025 2026 report\" — synthesizes numbers\nWebSearch: \"<market> growth rate CAGR billion\" — growth projections\nExtrapolation: TAM → SAM → SOM (Year 1)\n\n\n\nWrite research.md — write to docs/research.md in the current project directory. Create the directory if needed.\n\n\nOutput summary:\n\nKey findings (3-5 bullets)\nRecommendation: GO / NO-GO / PIVOT with brief reasoning\nPath to generated research.md\nSuggested next step: /validate <idea>"
      },
      {
        "title": "research.md Format",
        "body": "See references/research-template.md (bundled with this skill) for the full output template (frontmatter, 6 sections, tables)."
      },
      {
        "title": "Notes",
        "body": "Always use kebab-case for project directory names\nIf research.md already exists, ask before overwriting\nRun search queries in parallel when independent"
      },
      {
        "title": "MCP web_search not available",
        "body": "Cause: MCP server not running or not configured.\nFix: Use WebSearch/WebFetch as primary. For better results with engine routing (Reddit, GitHub, YouTube), set up SearXNG (private, self-hosted, free) and configure solograph MCP."
      },
      {
        "title": "Domain check returns wrong results",
        "body": "Cause: .app/.dev whois shows TLD creation date for unregistered domains.\nFix: Use the triple verification method (whois -> dig -> RDAP). Check Name Server and Registrar fields, not creation date."
      },
      {
        "title": "research.md already exists",
        "body": "Cause: Previous research run for this idea.\nFix: Skill asks before overwriting. Choose to merge new findings or start fresh."
      },
      {
        "title": "Reddit Deep Dive",
        "body": "MCP web_search or WebSearch — use for discovery (max 3 keywords for Reddit), get post URLs\nMCP Playwright — open old.reddit.com URLs to read full post + comments (bypasses CAPTCHA)\nExtract quotes — copy key phrases with attribution (u/username, subreddit, date)\nCross-post detection — same post in multiple subreddits = higher signal"
      },
      {
        "title": "Product Hunt Research",
        "body": "producthunt.com/visit-streaks — streak leaderboard (scrapeable via Playwright)\nproducthunt.com/@username — profile with social links, maker history, points\nPH API v2 is broken — redacts usernames/Twitter since Feb 2023, use scraping\nApify actors — check for DEPRECATED status before relying on them (mass deprecation Sep 2025)"
      },
      {
        "title": "TrustMRR Revenue Validation",
        "body": "trustmrr.com/startup/<slug> — Stripe-verified MRR, growth %, subscriptions, traffic\nWebFetch works — no auth needed, returns full page with JSON-LD structured data\nData fields: MRR, all-time revenue, last 30 days, active subs, tech stack, traffic (24h/7d/30d), category, founder X handle\nUse for: competitor revenue validation, market sizing with real data, tech stack discovery\nSearch: \"site:trustmrr.com <category or idea>\" to find similar startups with verified revenue\nApify scrapers: TrustMRR Scraper for bulk extraction"
      },
      {
        "title": "GitHub Library Discovery",
        "body": "MCP engines: github — often returns empty, use WebSearch as primary\ngithub.com/topics/<keyword> — browse topic pages via Playwright or WebFetch\nCheck stars, last update, open issues — avoid abandoned repos"
      },
      {
        "title": "Blocked Content Fallback Chain",
        "body": "MCP Playwright (best) → PullPush API (Reddit) → WebFetch → WebSearch snippets → MCP web_search include_raw_content\n\nIf a page returns 403/CAPTCHA via WebFetch:\n\nReddit: MCP Playwright → old.reddit.com (always works, no CAPTCHA)\nReddit search: PullPush API api.pullpush.io (structured JSON, full selftext)\nProduct Hunt / other sites: MCP Playwright browser_navigate (no captcha on most sites)\nGeneral: WebSearch snippets + WebSearch synthesis"
      }
    ],
    "body": "/research\n\nDeep research before PRD generation. Produces a structured research.md with competitive analysis, user pain points, SEO/ASO keywords, naming/domain options, and market sizing.\n\nMCP Tools (use if available)\n\nIf MCP tools are available, prefer them over CLI:\n\nkb_search(query, n_results) — search knowledge base for related docs\nweb_search(query, engines, include_raw_content) — web search with engine routing\nsession_search(query, project) — find how similar research was done before\nproject_info(name) — check project details and stacks\ncodegraph_explain(project) — architecture overview of an existing project (stack, patterns, deps)\ncodegraph_query(query) — raw Cypher queries against code graph (find shared packages, dependencies)\nproject_code_search(query, project) — semantic search over project source code\n\nMCP web_search supports engine override: engines=\"reddit\", engines=\"youtube\", etc. If MCP tools are not available, use WebSearch/WebFetch as primary. If MCP web_search tool is available, use it for better results.\n\nReddit Search Best Practices\nMax 3 keywords in reddit queries — more keywords = fewer results\nGood: \"product hunt outreach launch\" — Bad: \"product hunt scraper maker profiles linkedin outreach launch strategy\"\ninclude_raw_content=true rarely works for Reddit — use fallback chain below\nReddit Content Access — Fallback Chain\n\nWhen a search finds a relevant Reddit post, reading its full content requires a fallback chain:\n\n1. MCP Playwright (old.reddit.com)     ← BEST: bypasses CAPTCHA, full post + comments\n2. PullPush API (api.pullpush.io)      ← search by query/subreddit/author/score/date\n3. MCP web_search include_raw_content   ← sometimes works, often truncated\n4. WebFetch / WebSearch snippets        ← last resort, partial data only\n\n\nMethod 1: MCP Playwright (recommended for full post content)\n\nUse browser_navigate(\"https://old.reddit.com/r/...\") — old.reddit.com loads without CAPTCHA\nwww.reddit.com shows CAPTCHA (\"Prove your humanity\"), always use old.reddit.com\nSnapshot contains full post text + comments in structured YAML\nExample: old.reddit.com/r/indiehackers/comments/abc123/post_title/\n\nMethod 2: PullPush API (for search/discovery)\n\nEndpoint: https://api.pullpush.io/reddit/submission/search\nParams: q, subreddit, author, score (e.g. >10,<100), since/until (unix timestamps), size (max 100)\nRate limits: soft 15 req/min, hard 30 req/min, 1000 req/hr. Sleep 4 sec between requests.\nReturns JSON with full selftext, author, score, created_utc\nComment search: /reddit/comment/search (same params)\nCan use via curl:\ncurl -s \"https://api.pullpush.io/reddit/submission/search?q=product+hunt+launch&subreddit=indiehackers&size=10\"\n\n\nMethod 3: Reddit .json endpoint (often blocked)\n\nAppend .json to any Reddit URL: reddit.com/r/sub/comments/id.json\nReturns raw JSON with full post + comments\nFrequently blocked (403/429) — use as opportunistic fallback only\n\nMethod 4: PRAW (Reddit Official API, for live search/user profiles)\n\npraw-dev/praw — Python Reddit API Wrapper\nOAuth2 auth, built-in rate limiting, sync/async support\nBest for: live subreddit search, user profiles, comment trees\npip install praw / uv add praw\nSearch Strategy: Hybrid (MCP + WebSearch)\n\nUse multiple search backends together. Each has strengths:\n\nStep\tBest backend\tWhy\nCompetitors\tWebSearch + site:producthunt.com + site:g2.com\tBroad discovery + Product Hunt + B2B reviews\nReddit / Pain points\tMCP web_search with engines: reddit (max 3 keywords!) + MCP Playwright for full posts\tPullPush API, selftext in content\nYouTube reviews\tMCP web_search with engines: youtube\tVideo reviews (views = demand)\nMarket size\tWebSearch\tSynthesizes numbers from 10 sources\nSEO / ASO\tWebSearch\tBroader coverage, trend data\nPage scraping\tWebFetch or MCP web_search with include_raw_content\tUp to 5000 chars of page content\nHacker News\tWebSearch site:news.ycombinator.com\tHN discussions and opinions\nFunding / Companies\tWebSearch site:crunchbase.com\tCompetitor funding, team size\nVerified revenue\tWebFetch trustmrr.com/startup/<slug>\tStripe-verified MRR, growth, tech stack, traffic\nSearch Availability\n\nUse WebSearch/WebFetch as primary. If MCP web_search tool is available, use it for better results (supports engine routing and raw content extraction).\n\nSteps\n\nParse the idea from $ARGUMENTS. If empty, ask the user what idea they want to research.\n\nDetect product type — infer from the idea description:\n\nKeywords like \"app\", \"mobile\", \"iPhone\", \"Android\" → mobile (ios/android)\nKeywords like \"website\", \"SaaS\", \"dashboard\", \"web app\" → web\nKeywords like \"CLI\", \"terminal\", \"command line\" → cli\nKeywords like \"API\", \"backend\", \"service\" → api\nKeywords like \"extension\", \"plugin\", \"browser\" → web (extension)\nDefault if unclear → web\nOnly ask via AskUserQuestion if truly ambiguous (e.g., \"build a todo app\" could be web or mobile)\nThis determines which research sections apply (ASO for mobile, SEO for web, etc.)\n\nSearch knowledge base and past work:\n\nIf MCP kb_search available: kb_search(query=\"<idea keywords>\", n_results=5)\nIf MCP session_search available: session_search(query=\"<idea keywords>\") — check if this idea was researched before\nOtherwise: Grep for keywords in .md files\nCheck if research.md or prd.md already exist for this idea.\n\nCheck existing portfolio (if MCP codegraph tools available):\n\ncodegraph_explain(project=\"<similar project>\") — architecture overview of related projects in the portfolio\nproject_code_search(query=\"<relevant pattern>\", project=\"<sibling>\") — find reusable code, patterns, infrastructure\ncodegraph_query(\"MATCH (p:Project)-[:DEPENDS_ON]->(pkg:Package) WHERE pkg.name CONTAINS '<relevant tech>' RETURN p.name, pkg.name\") — find projects using similar tech\nThis helps assess: feasibility, reusable code, stack decisions, and time estimates\nIf no MCP tools available, skip this step.\n\nCompetitive analysis — use WebSearch (primary) + MCP web_search (if available):\n\n\"<idea> competitors alternatives 2026\" — broad discovery\n\"<idea> app review pricing\" — pricing data\nWebFetch or MCP include_raw_content=true: scrape competitor URLs for detailed pricing\nMCP engines: reddit or WebSearch: \"<idea> vs\" — user opinions\n\"site:producthunt.com <idea>\" — Product Hunt launches\n\"site:g2.com <idea>\" or \"site:capterra.com <idea>\" — B2B reviews\n\"site:crunchbase.com <competitor>\" — funding, team size\n\"site:trustmrr.com <idea>\" or WebFetch trustmrr.com/startup/<slug> — Stripe-verified MRR, growth %, tech stack, traffic (24h/7d/30d)\nFor each competitor extract: name, URL, pricing, key features, weaknesses, verified MRR (if on TrustMRR)\n\nUser pain points — use MCP web_search / WebSearch + YouTube:\n\nMCP engines: reddit or WebSearch: \"<problem>\" — Reddit discussions (max 3 keywords!)\nIf Reddit post found but content not available → open via MCP Playwright: browser_navigate(\"https://old.reddit.com/r/...\") — old.reddit.com bypasses CAPTCHA\nMCP engines: youtube or WebSearch: \"<problem> review\" — video reviews\n\"site:news.ycombinator.com <problem>\" — Hacker News opinions\nWebSearch: \"<problem> frustrating OR annoying\" — broader sweep\nSynthesis: top 5 pain points with quotes and source URLs\n\nSEO / ASO analysis (depends on product type from step 2):\n\nFor web apps:\n\n\"<competitor> SEO keywords ranking\" — competitor keywords\n\"<problem domain> search volume trends 2026\" — demand signals\nWebFetch or MCP include_raw_content: scrape competitor pages for meta tags\nResult: keyword table (keyword, intent, competition, relevance)\n\nFor mobile apps:\n\n\"<category> App Store top apps keywords 2026\" — category landscape\n\"site:reddit.com <competitor app> review\" — user complaints\nResult: ASO keywords, competitor ratings, common complaints\n\nNaming, domains, and company registration:\n\nGenerate 7-10 name candidates (mix of descriptive + invented/brandable)\nDomain availability: triple verification (whois → dig → RDAP)\nTrademark + company name conflict checks\n\nSee references/domain-check.md (bundled with this skill) for TLD priority tiers, bash scripts, gotchas, and trademark check methods.\n\nMarket sizing (TAM/SAM/SOM) — use WebSearch (primary):\n\nWebSearch: \"<market> market size 2025 2026 report\" — synthesizes numbers\nWebSearch: \"<market> growth rate CAGR billion\" — growth projections\nExtrapolation: TAM → SAM → SOM (Year 1)\n\nWrite research.md — write to docs/research.md in the current project directory. Create the directory if needed.\n\nOutput summary:\n\nKey findings (3-5 bullets)\nRecommendation: GO / NO-GO / PIVOT with brief reasoning\nPath to generated research.md\nSuggested next step: /validate <idea>\nresearch.md Format\n\nSee references/research-template.md (bundled with this skill) for the full output template (frontmatter, 6 sections, tables).\n\nNotes\nAlways use kebab-case for project directory names\nIf research.md already exists, ask before overwriting\nRun search queries in parallel when independent\nCommon Issues\nMCP web_search not available\n\nCause: MCP server not running or not configured. Fix: Use WebSearch/WebFetch as primary. For better results with engine routing (Reddit, GitHub, YouTube), set up SearXNG (private, self-hosted, free) and configure solograph MCP.\n\nDomain check returns wrong results\n\nCause: .app/.dev whois shows TLD creation date for unregistered domains. Fix: Use the triple verification method (whois -> dig -> RDAP). Check Name Server and Registrar fields, not creation date.\n\nresearch.md already exists\n\nCause: Previous research run for this idea. Fix: Skill asks before overwriting. Choose to merge new findings or start fresh.\n\nProactive Search Practices\nReddit Deep Dive\nMCP web_search or WebSearch — use for discovery (max 3 keywords for Reddit), get post URLs\nMCP Playwright — open old.reddit.com URLs to read full post + comments (bypasses CAPTCHA)\nExtract quotes — copy key phrases with attribution (u/username, subreddit, date)\nCross-post detection — same post in multiple subreddits = higher signal\nProduct Hunt Research\nproducthunt.com/visit-streaks — streak leaderboard (scrapeable via Playwright)\nproducthunt.com/@username — profile with social links, maker history, points\nPH API v2 is broken — redacts usernames/Twitter since Feb 2023, use scraping\nApify actors — check for DEPRECATED status before relying on them (mass deprecation Sep 2025)\nTrustMRR Revenue Validation\ntrustmrr.com/startup/<slug> — Stripe-verified MRR, growth %, subscriptions, traffic\nWebFetch works — no auth needed, returns full page with JSON-LD structured data\nData fields: MRR, all-time revenue, last 30 days, active subs, tech stack, traffic (24h/7d/30d), category, founder X handle\nUse for: competitor revenue validation, market sizing with real data, tech stack discovery\nSearch: \"site:trustmrr.com <category or idea>\" to find similar startups with verified revenue\nApify scrapers: TrustMRR Scraper for bulk extraction\nGitHub Library Discovery\nMCP engines: github — often returns empty, use WebSearch as primary\ngithub.com/topics/<keyword> — browse topic pages via Playwright or WebFetch\nCheck stars, last update, open issues — avoid abandoned repos\nBlocked Content Fallback Chain\nMCP Playwright (best) → PullPush API (Reddit) → WebFetch → WebSearch snippets → MCP web_search include_raw_content\n\n\nIf a page returns 403/CAPTCHA via WebFetch:\n\nReddit: MCP Playwright → old.reddit.com (always works, no CAPTCHA)\nReddit search: PullPush API api.pullpush.io (structured JSON, full selftext)\nProduct Hunt / other sites: MCP Playwright browser_navigate (no captcha on most sites)\nGeneral: WebSearch snippets + WebSearch synthesis"
  },
  "trust": {
    "sourceLabel": "tencent",
    "provenanceUrl": "https://clawhub.ai/fortunto2/solo-research",
    "publisherUrl": "https://clawhub.ai/fortunto2/solo-research",
    "owner": "fortunto2",
    "version": "1.7.1",
    "license": null,
    "verificationStatus": "Indexed source record"
  },
  "links": {
    "detailUrl": "https://openagent3.xyz/skills/solo-research",
    "downloadUrl": "https://openagent3.xyz/downloads/solo-research",
    "agentUrl": "https://openagent3.xyz/skills/solo-research/agent",
    "manifestUrl": "https://openagent3.xyz/skills/solo-research/agent.json",
    "briefUrl": "https://openagent3.xyz/skills/solo-research/agent.md"
  }
}