{
  "schemaVersion": "1.0",
  "item": {
    "slug": "offline-llama",
    "name": "Offline Llama",
    "source": "tencent",
    "type": "skill",
    "category": "AI 智能",
    "sourceUrl": "https://clawhub.ai/and-ray-m/offline-llama",
    "canonicalUrl": "https://clawhub.ai/and-ray-m/offline-llama",
    "targetPlatform": "OpenClaw"
  },
  "install": {
    "downloadMode": "redirect",
    "downloadUrl": "/downloads/offline-llama",
    "sourceDownloadUrl": "https://wry-manatee-359.convex.site/api/v1/download?slug=offline-llama",
    "sourcePlatform": "tencent",
    "targetPlatform": "OpenClaw",
    "installMethod": "Manual import",
    "extraction": "Extract archive",
    "prerequisites": [
      "OpenClaw"
    ],
    "packageFormat": "ZIP package",
    "includedAssets": [
      "SKILL.md"
    ],
    "primaryDoc": "SKILL.md",
    "quickSetup": [
      "Download the package from Yavira.",
      "Extract the archive and review SKILL.md first.",
      "Import or place the package into your OpenClaw setup."
    ],
    "agentAssist": {
      "summary": "Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.",
      "steps": [
        "Download the package from Yavira.",
        "Extract it into a folder your agent can access.",
        "Paste one of the prompts below and point your agent at the extracted folder."
      ],
      "prompts": [
        {
          "label": "New install",
          "body": "I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Tell me what you changed and call out any manual steps you could not complete."
        },
        {
          "label": "Upgrade existing",
          "body": "I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Summarize what changed and any follow-up checks I should run."
        }
      ]
    },
    "sourceHealth": {
      "source": "tencent",
      "status": "healthy",
      "reason": "direct_download_ok",
      "recommendedAction": "download",
      "checkedAt": "2026-04-30T16:55:25.780Z",
      "expiresAt": "2026-05-07T16:55:25.780Z",
      "httpStatus": 200,
      "finalUrl": "https://wry-manatee-359.convex.site/api/v1/download?slug=network",
      "contentType": "application/zip",
      "probeMethod": "head",
      "details": {
        "probeUrl": "https://wry-manatee-359.convex.site/api/v1/download?slug=network",
        "contentDisposition": "attachment; filename=\"network-1.0.0.zip\"",
        "redirectLocation": null,
        "bodySnippet": null
      },
      "scope": "source",
      "summary": "Source download looks usable.",
      "detail": "Yavira can redirect you to the upstream package for this source.",
      "primaryActionLabel": "Download for OpenClaw",
      "primaryActionHref": "/downloads/offline-llama"
    },
    "validation": {
      "installChecklist": [
        "Use the Yavira download entry.",
        "Review SKILL.md after the package is downloaded.",
        "Confirm the extracted package contains the expected setup assets."
      ],
      "postInstallChecks": [
        "Confirm the extracted package includes the expected docs or setup files.",
        "Validate the skill or prompts are available in your target agent workspace.",
        "Capture any manual follow-up steps the agent could not complete."
      ]
    },
    "downloadPageUrl": "https://openagent3.xyz/downloads/offline-llama",
    "agentPageUrl": "https://openagent3.xyz/skills/offline-llama/agent",
    "manifestUrl": "https://openagent3.xyz/skills/offline-llama/agent.json",
    "briefUrl": "https://openagent3.xyz/skills/offline-llama/agent.md"
  },
  "agentAssist": {
    "summary": "Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.",
    "steps": [
      "Download the package from Yavira.",
      "Extract it into a folder your agent can access.",
      "Paste one of the prompts below and point your agent at the extracted folder."
    ],
    "prompts": [
      {
        "label": "New install",
        "body": "I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Tell me what you changed and call out any manual steps you could not complete."
      },
      {
        "label": "Upgrade existing",
        "body": "I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Summarize what changed and any follow-up checks I should run."
      }
    ]
  },
  "documentation": {
    "source": "clawhub",
    "primaryDoc": "SKILL.md",
    "sections": [
      {
        "title": "offline-llama",
        "body": "Autonomously manage and use local Ollama models for continuous operation without internet dependency. Includes model health monitoring, automatic fallback, and self-healing capabilities."
      },
      {
        "title": "Overview",
        "body": "This skill enables autonomous operation with local Ollama models. It monitors model health, automatically switches between models when issues occur, and maintains functionality even without internet connectivity. The skill includes self-healing capabilities to restart services and clear resources when needed."
      },
      {
        "title": "Model Management",
        "body": "Health Monitoring: Continuously check model availability and performance\nAutomatic Fallback: Switch to alternative models when primary fails\nModel Switching: Dynamically select best available model for task"
      },
      {
        "title": "Self-Healing",
        "body": "Service Restart: Automatically restart Ollama when models become unavailable\nResource Management: Clear cache and temporary files to free resources\nModel Reinstallation: Reinstall problematic models automatically"
      },
      {
        "title": "Connectivity Awareness",
        "body": "Internet Detection: Monitor internet connectivity status\nSmart Fallback: Switch to remote models when local models unavailable and internet is present\nOffline Mode: Maintain full functionality without internet"
      },
      {
        "title": "Models",
        "body": "Primary: llama-3.1-8b-instruct (general tasks)\nSecondary: mistral-7b-instruct (faster responses)\nSpecialized: code-llama-7b (coding tasks)"
      },
      {
        "title": "Health Checks",
        "body": "Model Status: Monitor availability every 30 seconds\nLatency Tracking: Monitor response times every minute\nResource Usage: Monitor GPU/CPU and memory every 5 minutes"
      },
      {
        "title": "Fallback Strategies",
        "body": "Model Switching: Automatically switch to alternative local models\nResponse Retry: Retry failed requests with exponential backoff\nDegraded Mode: Continue with limited functionality if all models unavailable"
      },
      {
        "title": "When Internet is Available",
        "body": "Use local models primarily\nFallback to remote models if local models unavailable\nMaintain optimal performance"
      },
      {
        "title": "When Internet is Unavailable",
        "body": "Use local models exclusively\nContinue all operations without interruption\nProvide degraded functionality if needed"
      },
      {
        "title": "Model Management",
        "body": "model_status - Check current model health\nswitch_model - Manually switch between models\nrestart_ollama - Restart Ollama service"
      },
      {
        "title": "Health Monitoring",
        "body": "check_health - Run comprehensive health check\nmonitor_resources - Monitor system resources\nclear_cache - Clear model cache and temporary files"
      },
      {
        "title": "Automatic Actions",
        "body": "Service Restart: Triggered when model becomes unavailable\nResource Cleanup: Triggered when high memory usage detected\nModel Reinstallation: Triggered when persistent failures occur"
      },
      {
        "title": "Manual Intervention",
        "body": "Manual Restart: User can manually restart services\nCache Clearing: User can manually clear resources\nModel Updates: User can update models as needed"
      },
      {
        "title": "Security Considerations",
        "body": "All operations performed locally\nNo external dependencies required\nSecure model management\nPrivacy-preserving by default"
      },
      {
        "title": "Performance Optimization",
        "body": "Resource Monitoring: Track GPU/CPU usage and memory\nLatency Tracking: Monitor response times and performance\nModel Selection: Choose optimal model based on task requirements"
      },
      {
        "title": "Regular Tasks",
        "body": "Health Checks: Run periodic health checks\nCache Management: Clear unused cache regularly\nModel Updates: Keep models updated when possible"
      },
      {
        "title": "Troubleshooting",
        "body": "Log Analysis: Monitor logs for issues\nPerformance Metrics: Track performance over time\nError Handling: Graceful error handling and recovery"
      },
      {
        "title": "Integration",
        "body": "This skill integrates with:\n\nOllama: Local model management\nSystem Resources: Monitor and manage system resources\nNetwork: Detect internet connectivity\nOpenClaw: Seamless integration with existing tools"
      },
      {
        "title": "Future Enhancements",
        "body": "Model Training: Support for custom model training\nAdvanced Routing: Intelligent model selection based on task\nMulti-GPU Support: Scale across multiple GPUs\nCloud Sync: Optional cloud backup and synchronization"
      },
      {
        "title": "License",
        "body": "This skill is part of the OpenClaw ecosystem and follows the same licensing terms as OpenClaw itself."
      }
    ],
    "body": "offline-llama\n\nAutonomously manage and use local Ollama models for continuous operation without internet dependency. Includes model health monitoring, automatic fallback, and self-healing capabilities.\n\nOverview\n\nThis skill enables autonomous operation with local Ollama models. It monitors model health, automatically switches between models when issues occur, and maintains functionality even without internet connectivity. The skill includes self-healing capabilities to restart services and clear resources when needed.\n\nCore Features\nModel Management\nHealth Monitoring: Continuously check model availability and performance\nAutomatic Fallback: Switch to alternative models when primary fails\nModel Switching: Dynamically select best available model for task\nSelf-Healing\nService Restart: Automatically restart Ollama when models become unavailable\nResource Management: Clear cache and temporary files to free resources\nModel Reinstallation: Reinstall problematic models automatically\nConnectivity Awareness\nInternet Detection: Monitor internet connectivity status\nSmart Fallback: Switch to remote models when local models unavailable and internet is present\nOffline Mode: Maintain full functionality without internet\nConfiguration\nModels\nPrimary: llama-3.1-8b-instruct (general tasks)\nSecondary: mistral-7b-instruct (faster responses)\nSpecialized: code-llama-7b (coding tasks)\nHealth Checks\nModel Status: Monitor availability every 30 seconds\nLatency Tracking: Monitor response times every minute\nResource Usage: Monitor GPU/CPU and memory every 5 minutes\nFallback Strategies\nModel Switching: Automatically switch to alternative local models\nResponse Retry: Retry failed requests with exponential backoff\nDegraded Mode: Continue with limited functionality if all models unavailable\nUsage\nWhen Internet is Available\nUse local models primarily\nFallback to remote models if local models unavailable\nMaintain optimal performance\nWhen Internet is Unavailable\nUse local models exclusively\nContinue all operations without interruption\nProvide degraded functionality if needed\nCommands\nModel Management\nmodel_status - Check current model health\nswitch_model - Manually switch between models\nrestart_ollama - Restart Ollama service\nHealth Monitoring\ncheck_health - Run comprehensive health check\nmonitor_resources - Monitor system resources\nclear_cache - Clear model cache and temporary files\nSelf-Healing\nAutomatic Actions\nService Restart: Triggered when model becomes unavailable\nResource Cleanup: Triggered when high memory usage detected\nModel Reinstallation: Triggered when persistent failures occur\nManual Intervention\nManual Restart: User can manually restart services\nCache Clearing: User can manually clear resources\nModel Updates: User can update models as needed\nSecurity Considerations\nAll operations performed locally\nNo external dependencies required\nSecure model management\nPrivacy-preserving by default\nPerformance Optimization\nResource Monitoring: Track GPU/CPU usage and memory\nLatency Tracking: Monitor response times and performance\nModel Selection: Choose optimal model based on task requirements\nMaintenance\nRegular Tasks\nHealth Checks: Run periodic health checks\nCache Management: Clear unused cache regularly\nModel Updates: Keep models updated when possible\nTroubleshooting\nLog Analysis: Monitor logs for issues\nPerformance Metrics: Track performance over time\nError Handling: Graceful error handling and recovery\nIntegration\n\nThis skill integrates with:\n\nOllama: Local model management\nSystem Resources: Monitor and manage system resources\nNetwork: Detect internet connectivity\nOpenClaw: Seamless integration with existing tools\nFuture Enhancements\nModel Training: Support for custom model training\nAdvanced Routing: Intelligent model selection based on task\nMulti-GPU Support: Scale across multiple GPUs\nCloud Sync: Optional cloud backup and synchronization\nLicense\n\nThis skill is part of the OpenClaw ecosystem and follows the same licensing terms as OpenClaw itself."
  },
  "trust": {
    "sourceLabel": "tencent",
    "provenanceUrl": "https://clawhub.ai/and-ray-m/offline-llama",
    "publisherUrl": "https://clawhub.ai/and-ray-m/offline-llama",
    "owner": "and-ray-m",
    "version": "1.0.0",
    "license": null,
    "verificationStatus": "Indexed source record"
  },
  "links": {
    "detailUrl": "https://openagent3.xyz/skills/offline-llama",
    "downloadUrl": "https://openagent3.xyz/downloads/offline-llama",
    "agentUrl": "https://openagent3.xyz/skills/offline-llama/agent",
    "manifestUrl": "https://openagent3.xyz/skills/offline-llama/agent.json",
    "briefUrl": "https://openagent3.xyz/skills/offline-llama/agent.md"
  }
}