{
  "schemaVersion": "1.0",
  "item": {
    "slug": "anakin",
    "name": "anakin",
    "source": "tencent",
    "type": "skill",
    "category": "开发工具",
    "sourceUrl": "https://clawhub.ai/Viraal-Bambori/anakin",
    "canonicalUrl": "https://clawhub.ai/Viraal-Bambori/anakin",
    "targetPlatform": "OpenClaw"
  },
  "install": {
    "downloadMode": "redirect",
    "downloadUrl": "/downloads/anakin",
    "sourceDownloadUrl": "https://wry-manatee-359.convex.site/api/v1/download?slug=anakin",
    "sourcePlatform": "tencent",
    "targetPlatform": "OpenClaw",
    "installMethod": "Manual import",
    "extraction": "Extract archive",
    "prerequisites": [
      "OpenClaw"
    ],
    "packageFormat": "ZIP package",
    "includedAssets": [
      "CHANGELOG.md",
      "README.md",
      "SKILL.md",
      "skill.json",
      "rules/install.md"
    ],
    "primaryDoc": "SKILL.md",
    "quickSetup": [
      "Download the package from Yavira.",
      "Extract the archive and review SKILL.md first.",
      "Import or place the package into your OpenClaw setup."
    ],
    "agentAssist": {
      "summary": "Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.",
      "steps": [
        "Download the package from Yavira.",
        "Extract it into a folder your agent can access.",
        "Paste one of the prompts below and point your agent at the extracted folder."
      ],
      "prompts": [
        {
          "label": "New install",
          "body": "I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Then review README.md for any prerequisites, environment setup, or post-install checks. Tell me what you changed and call out any manual steps you could not complete."
        },
        {
          "label": "Upgrade existing",
          "body": "I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Then review README.md for any prerequisites, environment setup, or post-install checks. Summarize what changed and any follow-up checks I should run."
        }
      ]
    },
    "sourceHealth": {
      "source": "tencent",
      "status": "healthy",
      "reason": "direct_download_ok",
      "recommendedAction": "download",
      "checkedAt": "2026-05-07T17:22:31.273Z",
      "expiresAt": "2026-05-14T17:22:31.273Z",
      "httpStatus": 200,
      "finalUrl": "https://wry-manatee-359.convex.site/api/v1/download?slug=afrexai-annual-report",
      "contentType": "application/zip",
      "probeMethod": "head",
      "details": {
        "probeUrl": "https://wry-manatee-359.convex.site/api/v1/download?slug=afrexai-annual-report",
        "contentDisposition": "attachment; filename=\"afrexai-annual-report-1.0.0.zip\"",
        "redirectLocation": null,
        "bodySnippet": null
      },
      "scope": "source",
      "summary": "Source download looks usable.",
      "detail": "Yavira can redirect you to the upstream package for this source.",
      "primaryActionLabel": "Download for OpenClaw",
      "primaryActionHref": "/downloads/anakin"
    },
    "validation": {
      "installChecklist": [
        "Use the Yavira download entry.",
        "Review SKILL.md after the package is downloaded.",
        "Confirm the extracted package contains the expected setup assets."
      ],
      "postInstallChecks": [
        "Confirm the extracted package includes the expected docs or setup files.",
        "Validate the skill or prompts are available in your target agent workspace.",
        "Capture any manual follow-up steps the agent could not complete."
      ]
    },
    "downloadPageUrl": "https://openagent3.xyz/downloads/anakin",
    "agentPageUrl": "https://openagent3.xyz/skills/anakin/agent",
    "manifestUrl": "https://openagent3.xyz/skills/anakin/agent.json",
    "briefUrl": "https://openagent3.xyz/skills/anakin/agent.md"
  },
  "agentAssist": {
    "summary": "Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.",
    "steps": [
      "Download the package from Yavira.",
      "Extract it into a folder your agent can access.",
      "Paste one of the prompts below and point your agent at the extracted folder."
    ],
    "prompts": [
      {
        "label": "New install",
        "body": "I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Then review README.md for any prerequisites, environment setup, or post-install checks. Tell me what you changed and call out any manual steps you could not complete."
      },
      {
        "label": "Upgrade existing",
        "body": "I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Then review README.md for any prerequisites, environment setup, or post-install checks. Summarize what changed and any follow-up checks I should run."
      }
    ]
  },
  "documentation": {
    "source": "clawhub",
    "primaryDoc": "SKILL.md",
    "sections": [
      {
        "title": "Anakin - Web Data Extraction",
        "body": "Convert websites into clean data at scale using the anakin-cli. Supports single URL scraping, batch scraping, AI-powered search, and autonomous deep research."
      },
      {
        "title": "Installation & Authentication",
        "body": "Check status and authentication:\n\nanakin status\n\nOutput when ready:\n\n✓ Authenticated\nEndpoint: https://api.anakin.io\nAccount: user@example.com\n\nIf not installed: pip install anakin-cli\n\nAlways refer to the installation rules in rules/install.md for more information if the user is not logged in.\n\nIf not authenticated, run:\n\nanakin login --api-key \"ak-your-key-here\"\n\nGet your API key from anakin.io/dashboard."
      },
      {
        "title": "Organization",
        "body": "Create a .anakin/ folder in the working directory unless it already exists to store results. Add .anakin/ to the .gitignore file if not already there. Always use -o to write directly to file (avoids flooding context):\n\nmkdir -p .anakin\necho \".anakin/\" >> .gitignore\nanakin scrape \"https://example.com\" -o .anakin/output.md"
      },
      {
        "title": "1. Scrape a Single URL",
        "body": "Extract content from a single web page in multiple formats.\n\nWhen to use:\n\nExtracting content from a single web page\nConverting a webpage to clean markdown\nExtracting structured data from one URL\nGetting full raw API response with metadata\n\nBasic usage:\n\n# Clean readable text (default markdown format)\nanakin scrape \"https://example.com\" -o output.md\n\n# Structured data (JSON)\nanakin scrape \"https://example.com\" --format json -o output.json\n\n# Full API response with HTML and metadata\nanakin scrape \"https://example.com\" --format raw -o output.json\n\nAdvanced options:\n\n# JavaScript-heavy or single-page app sites\nanakin scrape \"https://example.com\" --browser -o output.md\n\n# Geo-targeted scraping (country code)\nanakin scrape \"https://example.com\" --country gb -o output.md\n\n# Custom timeout for slow pages (in seconds)\nanakin scrape \"https://example.com\" --timeout 300 -o output.md"
      },
      {
        "title": "2. Batch Scrape Multiple URLs",
        "body": "Scrape up to 10 URLs at once for efficient parallel processing.\n\nWhen to use:\n\nScraping multiple web pages simultaneously\nComparing products across different sites\nCollecting multiple articles or pages\nGathering data from several sources at once\n\nBasic usage:\n\n# Batch scrape multiple URLs (up to 10)\nanakin scrape-batch \"https://example.com/page1\" \"https://example.com/page2\" \"https://example.com/page3\" -o batch-results.json\n\nFor large lists (>10 URLs):\n\n# First batch (URLs 1-10)\nanakin scrape-batch \"https://url1.com\" ... \"https://url10.com\" -o batch-1.json\n\n# Second batch (URLs 11-20)\nanakin scrape-batch \"https://url11.com\" ... \"https://url20.com\" -o batch-2.json\n\nOutput format: JSON file with combined results, each URL's status (success/failure), content, metadata, and any errors."
      },
      {
        "title": "3. AI-Powered Web Search",
        "body": "Run intelligent web searches to find pages, answer questions, and discover sources.\n\nWhen to use:\n\nFinding pages on a specific topic\nAnswering questions with web sources\nDiscovering relevant sources for research\nGathering links before scraping specific pages\nQuick factual lookups\n\nBasic usage:\n\n# AI-powered web search\nanakin search \"your search query here\" -o search-results.json\n\nFollow-up workflow:\n\n# 1. Search for relevant pages\nanakin search \"machine learning tutorials\" -o search-results.json\n\n# 2. Scrape specific results for full content\nanakin scrape \"https://result-url-from-search.com\" -o page.md\n\nOutput format: JSON file with search results including titles, URLs, snippets, relevance scores, and metadata."
      },
      {
        "title": "4. Deep Agentic Research",
        "body": "Run comprehensive autonomous research that explores the web and returns detailed reports.\n\nWhen to use:\n\nComprehensive research on complex topics\nMarket analysis requiring multiple sources\nTechnical deep-dives across documentation and articles\nComparison research (products, technologies, approaches)\nQuestions requiring synthesis from many sources\n\nBasic usage:\n\n# Deep agentic research (takes 1-5 minutes)\nanakin research \"your research topic or question\" -o research-report.json\n\n# With extended timeout for complex topics\nanakin research \"comprehensive analysis of quantum computing\" --timeout 600 -o research-report.json\n\n⏱️ Important: Deep research takes 1-5 minutes and runs autonomously. Always inform the user about this duration before starting.\n\nWhat it does:\n\nAutonomously searches for relevant sources\nScrapes and analyzes multiple pages\nSynthesizes information across sources\nGenerates comprehensive reports with citations\nProvides key insights and conclusions\n\nOutput format: JSON file with executive summary, detailed report by subtopics, key insights, citations with URLs, confidence scores, and related topics."
      },
      {
        "title": "Decision Guide",
        "body": "Use anakin scrape when:\n\nYou have a single specific URL to extract\nYou need content in markdown, JSON, or raw format\nThe page is static or JavaScript-heavy (use --browser)\n\nUse anakin scrape-batch when:\n\nYou have 2-10 URLs to scrape simultaneously\nYou need efficient parallel processing\nYou want combined results in one file\n\nUse anakin search when:\n\nYou need to find relevant URLs first\nYou want quick factual lookups\nYou need results in under 30 seconds\nYou know what you're looking for\n\nUse anakin research when:\n\nYou need comprehensive analysis across 5+ sources\nThe topic is complex and requires deep exploration\nYou want a synthesized report with insights\nYou can wait 1-5 minutes for autonomous research\nThe question requires comparing multiple perspectives"
      },
      {
        "title": "URL Handling",
        "body": "Always quote URLs to prevent shell interpretation of ?, &, # characters\nExample: anakin scrape \"https://example.com?param=value\" not anakin scrape https://example.com?param=value"
      },
      {
        "title": "Output Management",
        "body": "Always use -o <file> to save output to a file rather than flooding the terminal\nChoose appropriate output filenames based on content type"
      },
      {
        "title": "Format Selection",
        "body": "Default to markdown for readability unless user explicitly asks for JSON or raw\nUse --format json for structured data processing\nUse --format raw for full API response with HTML"
      },
      {
        "title": "Special Cases",
        "body": "Use --browser only when standard scrape returns empty or incomplete content\nFor batch scraping: Maximum 10 URLs per command — split larger lists\nFor research: Always warn about 1-5 minute duration before starting"
      },
      {
        "title": "Rate Limiting",
        "body": "On HTTP 429 errors (rate limit), wait before retrying\nDo not loop immediately on rate limit errors"
      },
      {
        "title": "Authentication",
        "body": "On HTTP 401 errors, re-run anakin login rather than retrying the same command"
      },
      {
        "title": "Error Handling",
        "body": "ErrorSolutionHTTP 401 (Unauthorized)Re-run anakin login --api-key \"your-key\"HTTP 429 (Rate Limited)Wait before retrying, do not loop immediatelyEmpty contentTry adding --browser flag for JavaScript-heavy sitesTimeoutIncrease with --timeout <seconds> for slow pagesBatch partial failureCheck output JSON for individual statuses, retry failed URLs with --browserResearch failsFall back to search + multiple scrape calls manually"
      },
      {
        "title": "Markdown (default for scrape)",
        "body": "Clean, readable text stripped of navigation and ads\nBest for human reading and summarization\nFile extension: .md"
      },
      {
        "title": "JSON (structured)",
        "body": "Structured data with title, content, metadata\nBest for processing and parsing\nFile extension: .json"
      },
      {
        "title": "Raw (full response)",
        "body": "Full API response including HTML, links, images, metadata\nBest for debugging or accessing all available data\nFile extension: .json"
      },
      {
        "title": "Example 1: Article extraction",
        "body": "anakin scrape \"https://blog.example.com/article\" -o article.md"
      },
      {
        "title": "Example 2: Product comparison",
        "body": "anakin scrape-batch \"https://store1.com/product\" \"https://store2.com/product\" \"https://store3.com/product\" -o products.json"
      },
      {
        "title": "Example 3: Find and scrape",
        "body": "# Step 1: Find relevant URLs\nanakin search \"best coffee shops in Seattle\" -o coffee-search.json\n\n# Step 2: Scrape the top results\nanakin scrape-batch \"url1\" \"url2\" \"url3\" -o coffee-details.json"
      },
      {
        "title": "Example 4: Market research",
        "body": "anakin research \"market trends in electric vehicle adoption 2024-2026\" -o ev-research.json"
      },
      {
        "title": "Example 5: JavaScript-heavy site",
        "body": "anakin scrape \"https://spa-application.com\" --browser -o spa-content.md"
      },
      {
        "title": "Example 6: Geo-targeted content",
        "body": "anakin scrape \"https://news-site.com\" --country us -o us-news.md\nanakin scrape \"https://news-site.com\" --country gb -o gb-news.md"
      },
      {
        "title": "Best Practices",
        "body": "Start simple: Try basic scrape first, add flags only if needed\nBe specific: Use clear, specific search queries and research topics\nQuote URLs: Always wrap URLs in quotes\nSave output: Always use -o flag to save results to files\nCheck status: Run anakin status before starting work\nBatch wisely: Group similar URLs together, max 10 per batch\nWait on rate limits: Don't retry immediately on 429 errors\nChoose the right tool:\n\nSingle page → scrape\nMultiple pages → scrape-batch\nDon't have URLs → search first\nNeed deep analysis → research"
      },
      {
        "title": "Authentication issues",
        "body": "# Check status\nanakin status\n\n# Re-authenticate\nanakin login --api-key \"ak-your-key-here\""
      },
      {
        "title": "Empty or incomplete content",
        "body": "Add --browser flag for JavaScript-heavy sites\nIncrease timeout with --timeout 300\nCheck if the site requires specific geo-location with --country <code>"
      },
      {
        "title": "Rate limiting",
        "body": "Wait before retrying (don't loop immediately)\nConsider spacing out requests for large batch operations\nCheck your API plan limits at anakin.io/dashboard"
      },
      {
        "title": "Resources",
        "body": "Anakin Website\nAnakin Dashboard - Get API keys and check usage\nanakin-cli on PyPI\nSupport"
      }
    ],
    "body": "Anakin - Web Data Extraction\n\nConvert websites into clean data at scale using the anakin-cli. Supports single URL scraping, batch scraping, AI-powered search, and autonomous deep research.\n\nInstallation & Authentication\n\nCheck status and authentication:\n\nanakin status\n\n\nOutput when ready:\n\n✓ Authenticated\nEndpoint: https://api.anakin.io\nAccount: user@example.com\n\n\nIf not installed: pip install anakin-cli\n\nAlways refer to the installation rules in rules/install.md for more information if the user is not logged in.\n\nIf not authenticated, run:\n\nanakin login --api-key \"ak-your-key-here\"\n\n\nGet your API key from anakin.io/dashboard.\n\nOrganization\n\nCreate a .anakin/ folder in the working directory unless it already exists to store results. Add .anakin/ to the .gitignore file if not already there. Always use -o to write directly to file (avoids flooding context):\n\nmkdir -p .anakin\necho \".anakin/\" >> .gitignore\nanakin scrape \"https://example.com\" -o .anakin/output.md\n\nCapabilities\n1. Scrape a Single URL\n\nExtract content from a single web page in multiple formats.\n\nWhen to use:\n\nExtracting content from a single web page\nConverting a webpage to clean markdown\nExtracting structured data from one URL\nGetting full raw API response with metadata\n\nBasic usage:\n\n# Clean readable text (default markdown format)\nanakin scrape \"https://example.com\" -o output.md\n\n# Structured data (JSON)\nanakin scrape \"https://example.com\" --format json -o output.json\n\n# Full API response with HTML and metadata\nanakin scrape \"https://example.com\" --format raw -o output.json\n\n\nAdvanced options:\n\n# JavaScript-heavy or single-page app sites\nanakin scrape \"https://example.com\" --browser -o output.md\n\n# Geo-targeted scraping (country code)\nanakin scrape \"https://example.com\" --country gb -o output.md\n\n# Custom timeout for slow pages (in seconds)\nanakin scrape \"https://example.com\" --timeout 300 -o output.md\n\n2. Batch Scrape Multiple URLs\n\nScrape up to 10 URLs at once for efficient parallel processing.\n\nWhen to use:\n\nScraping multiple web pages simultaneously\nComparing products across different sites\nCollecting multiple articles or pages\nGathering data from several sources at once\n\nBasic usage:\n\n# Batch scrape multiple URLs (up to 10)\nanakin scrape-batch \"https://example.com/page1\" \"https://example.com/page2\" \"https://example.com/page3\" -o batch-results.json\n\n\nFor large lists (>10 URLs):\n\n# First batch (URLs 1-10)\nanakin scrape-batch \"https://url1.com\" ... \"https://url10.com\" -o batch-1.json\n\n# Second batch (URLs 11-20)\nanakin scrape-batch \"https://url11.com\" ... \"https://url20.com\" -o batch-2.json\n\n\nOutput format: JSON file with combined results, each URL's status (success/failure), content, metadata, and any errors.\n\n3. AI-Powered Web Search\n\nRun intelligent web searches to find pages, answer questions, and discover sources.\n\nWhen to use:\n\nFinding pages on a specific topic\nAnswering questions with web sources\nDiscovering relevant sources for research\nGathering links before scraping specific pages\nQuick factual lookups\n\nBasic usage:\n\n# AI-powered web search\nanakin search \"your search query here\" -o search-results.json\n\n\nFollow-up workflow:\n\n# 1. Search for relevant pages\nanakin search \"machine learning tutorials\" -o search-results.json\n\n# 2. Scrape specific results for full content\nanakin scrape \"https://result-url-from-search.com\" -o page.md\n\n\nOutput format: JSON file with search results including titles, URLs, snippets, relevance scores, and metadata.\n\n4. Deep Agentic Research\n\nRun comprehensive autonomous research that explores the web and returns detailed reports.\n\nWhen to use:\n\nComprehensive research on complex topics\nMarket analysis requiring multiple sources\nTechnical deep-dives across documentation and articles\nComparison research (products, technologies, approaches)\nQuestions requiring synthesis from many sources\n\nBasic usage:\n\n# Deep agentic research (takes 1-5 minutes)\nanakin research \"your research topic or question\" -o research-report.json\n\n# With extended timeout for complex topics\nanakin research \"comprehensive analysis of quantum computing\" --timeout 600 -o research-report.json\n\n\n⏱️ Important: Deep research takes 1-5 minutes and runs autonomously. Always inform the user about this duration before starting.\n\nWhat it does:\n\nAutonomously searches for relevant sources\nScrapes and analyzes multiple pages\nSynthesizes information across sources\nGenerates comprehensive reports with citations\nProvides key insights and conclusions\n\nOutput format: JSON file with executive summary, detailed report by subtopics, key insights, citations with URLs, confidence scores, and related topics.\n\nDecision Guide\n\nUse anakin scrape when:\n\nYou have a single specific URL to extract\nYou need content in markdown, JSON, or raw format\nThe page is static or JavaScript-heavy (use --browser)\n\nUse anakin scrape-batch when:\n\nYou have 2-10 URLs to scrape simultaneously\nYou need efficient parallel processing\nYou want combined results in one file\n\nUse anakin search when:\n\nYou need to find relevant URLs first\nYou want quick factual lookups\nYou need results in under 30 seconds\nYou know what you're looking for\n\nUse anakin research when:\n\nYou need comprehensive analysis across 5+ sources\nThe topic is complex and requires deep exploration\nYou want a synthesized report with insights\nYou can wait 1-5 minutes for autonomous research\nThe question requires comparing multiple perspectives\nGuardrails\nURL Handling\nAlways quote URLs to prevent shell interpretation of ?, &, # characters\nExample: anakin scrape \"https://example.com?param=value\" not anakin scrape https://example.com?param=value\nOutput Management\nAlways use -o <file> to save output to a file rather than flooding the terminal\nChoose appropriate output filenames based on content type\nFormat Selection\nDefault to markdown for readability unless user explicitly asks for JSON or raw\nUse --format json for structured data processing\nUse --format raw for full API response with HTML\nSpecial Cases\nUse --browser only when standard scrape returns empty or incomplete content\nFor batch scraping: Maximum 10 URLs per command — split larger lists\nFor research: Always warn about 1-5 minute duration before starting\nRate Limiting\nOn HTTP 429 errors (rate limit), wait before retrying\nDo not loop immediately on rate limit errors\nAuthentication\nOn HTTP 401 errors, re-run anakin login rather than retrying the same command\nError Handling\nError\tSolution\nHTTP 401 (Unauthorized)\tRe-run anakin login --api-key \"your-key\"\nHTTP 429 (Rate Limited)\tWait before retrying, do not loop immediately\nEmpty content\tTry adding --browser flag for JavaScript-heavy sites\nTimeout\tIncrease with --timeout <seconds> for slow pages\nBatch partial failure\tCheck output JSON for individual statuses, retry failed URLs with --browser\nResearch fails\tFall back to search + multiple scrape calls manually\nOutput Formats\nMarkdown (default for scrape)\nClean, readable text stripped of navigation and ads\nBest for human reading and summarization\nFile extension: .md\nJSON (structured)\nStructured data with title, content, metadata\nBest for processing and parsing\nFile extension: .json\nRaw (full response)\nFull API response including HTML, links, images, metadata\nBest for debugging or accessing all available data\nFile extension: .json\nExamples\nExample 1: Article extraction\nanakin scrape \"https://blog.example.com/article\" -o article.md\n\nExample 2: Product comparison\nanakin scrape-batch \"https://store1.com/product\" \"https://store2.com/product\" \"https://store3.com/product\" -o products.json\n\nExample 3: Find and scrape\n# Step 1: Find relevant URLs\nanakin search \"best coffee shops in Seattle\" -o coffee-search.json\n\n# Step 2: Scrape the top results\nanakin scrape-batch \"url1\" \"url2\" \"url3\" -o coffee-details.json\n\nExample 4: Market research\nanakin research \"market trends in electric vehicle adoption 2024-2026\" -o ev-research.json\n\nExample 5: JavaScript-heavy site\nanakin scrape \"https://spa-application.com\" --browser -o spa-content.md\n\nExample 6: Geo-targeted content\nanakin scrape \"https://news-site.com\" --country us -o us-news.md\nanakin scrape \"https://news-site.com\" --country gb -o gb-news.md\n\nBest Practices\nStart simple: Try basic scrape first, add flags only if needed\nBe specific: Use clear, specific search queries and research topics\nQuote URLs: Always wrap URLs in quotes\nSave output: Always use -o flag to save results to files\nCheck status: Run anakin status before starting work\nBatch wisely: Group similar URLs together, max 10 per batch\nWait on rate limits: Don't retry immediately on 429 errors\nChoose the right tool:\nSingle page → scrape\nMultiple pages → scrape-batch\nDon't have URLs → search first\nNeed deep analysis → research\nTroubleshooting\nAuthentication issues\n# Check status\nanakin status\n\n# Re-authenticate\nanakin login --api-key \"ak-your-key-here\"\n\nEmpty or incomplete content\nAdd --browser flag for JavaScript-heavy sites\nIncrease timeout with --timeout 300\nCheck if the site requires specific geo-location with --country <code>\nRate limiting\nWait before retrying (don't loop immediately)\nConsider spacing out requests for large batch operations\nCheck your API plan limits at anakin.io/dashboard\nResources\nAnakin Website\nAnakin Dashboard - Get API keys and check usage\nanakin-cli on PyPI\nSupport"
  },
  "trust": {
    "sourceLabel": "tencent",
    "provenanceUrl": "https://clawhub.ai/Viraal-Bambori/anakin",
    "publisherUrl": "https://clawhub.ai/Viraal-Bambori/anakin",
    "owner": "Viraal-Bambori",
    "version": "1.0.0",
    "license": null,
    "verificationStatus": "Indexed source record"
  },
  "links": {
    "detailUrl": "https://openagent3.xyz/skills/anakin",
    "downloadUrl": "https://openagent3.xyz/downloads/anakin",
    "agentUrl": "https://openagent3.xyz/skills/anakin/agent",
    "manifestUrl": "https://openagent3.xyz/skills/anakin/agent.json",
    "briefUrl": "https://openagent3.xyz/skills/anakin/agent.md"
  }
}