Requirements
- Target platform
- OpenClaw
- Install method
- Manual import
- Extraction
- Extract archive
- Prerequisites
- OpenClaw
- Primary doc
- SKILL.md
Proactive self-monitoring of infrastructure, services, and health. Tracks disk/memory/load, service health, cron job status, recent errors. Auto-fixes safe i...
Proactive self-monitoring of infrastructure, services, and health. Tracks disk/memory/load, service health, cron job status, recent errors. Auto-fixes safe i...
Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.
I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Then review README.md for any prerequisites, environment setup, or post-install checks. Tell me what you changed and call out any manual steps you could not complete.
I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Then review README.md for any prerequisites, environment setup, or post-install checks. Summarize what changed and any follow-up checks I should run.
Proactive self-monitoring: infrastructure, services, and health.
Run during heartbeats or scheduled checks.
# Disk usage df -h / | awk 'NR==2 {print $5}' | tr -d '%' # Memory usage free -m | awk 'NR==2 {printf "%.0f", $3/$2*100}' # Load average uptime | awk -F'load average:' '{print $2}' | awk -F',' '{print $1}' # Top processes by memory ps aux --sort=-%mem | head -10 # Top processes by CPU ps aux --sort=-%cpu | head -10 Thresholds: MetricWarningCriticalDisk> 80%> 90%Memory> 85%> 95%Load> 2.0> 4.0
# Check if a process is running pgrep -f "your_process_name" >/dev/null && echo "OK" || echo "FAIL" # Check HTTP endpoint curl -s -o /dev/null -w "%{http_code}" http://localhost:8080/health # Check systemd service systemctl is-active --quiet nginx && echo "OK" || echo "FAIL" # Check Docker container docker ps --filter "name=mycontainer" --filter "status=running" -q | grep -q . && echo "OK" || echo "FAIL" # Tailscale (if using) tailscale status --json 2>/dev/null | jq -r '.Self.Online' || echo "FAIL"
# Check recent cron executions grep CRON /var/log/syslog | tail -20 # Count failures in last 24h grep -c "CRON.*error\|CRON.*fail" /var/log/syslog # List scheduled jobs crontab -l
# Check system logs for errors journalctl -p err --since "1 hour ago" 2>/dev/null | tail -20 # Check application logs tail -50 ~/workspace/projects/*/logs/*.log 2>/dev/null | grep -i "error" # Check dmesg for hardware/kernel issues dmesg | tail -20 | grep -i "error\|fail\|warn"
#!/bin/bash # Quick health snapshot DISK=$(df -h / | awk 'NR==2 {print $5}' | tr -d '%') MEM=$(free -m | awk 'NR==2 {printf "%.0f", $3/$2*100}') LOAD=$(uptime | awk -F'load average:' '{print $2}' | awk -F',' '{print $1}' | xargs) echo "Disk: ${DISK}% | Mem: ${MEM}% | Load: ${LOAD}" # Alert if thresholds exceeded [ "$DISK" -gt 90 ] && echo "β οΈ Disk critical!" [ "$MEM" -gt 95 ] && echo "β οΈ Memory critical!"
When issues detected: IssueAuto-ActionAlert?Disk > 90%Clean temp files, old logsYesKey process downAttempt restartYesCron 3+ failuresGenerate reportYesMemory > 95%List top processesYes Auto-fixable (safe): # Clean old logs (> 7 days) find /var/log -name "*.log" -mtime +7 -delete 2>/dev/null find ~/.cache -type f -mtime +7 -delete 2>/dev/null # Clean temp files rm -f /tmp/agent-temp-* 2>/dev/null rm -rf ~/.cache/pip 2>/dev/null
Add to your crontab or task scheduler: # Run health check every 30 minutes */30 * * * * /path/to/health-check.sh >> /var/log/health-check.log 2>&1 Or run manually as part of your workflow: ./health-check.sh
Agent frameworks, memory systems, reasoning layers, and model-native orchestration.
Largest current source with strong distribution and engagement signals.