Requirements
- Target platform
- OpenClaw
- Install method
- Manual import
- Extraction
- Extract archive
- Prerequisites
- OpenClaw
- Primary doc
- SKILL.md
Audit your AI agent setup for performance, cost, and ROI. Scans OpenClaw config, cron jobs, session history, and model usage to find waste and recommend opti...
Audit your AI agent setup for performance, cost, and ROI. Scans OpenClaw config, cron jobs, session history, and model usage to find waste and recommend opti...
Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.
I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Tell me what you changed and call out any manual steps you could not complete.
I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Summarize what changed and any follow-up checks I should run.
Scan your entire OpenClaw setup and get actionable cost/performance recommendations.
Scans config โ reads OpenClaw config to map models to agents/tasks Analyzes cron history โ checks every cron job's model, token usage, runtime, success rate Classifies tasks โ determines complexity level of each task Calculates costs โ per agent, per cron, per task type using provider pricing Recommends changes โ with confidence levels and risk warnings Generates report โ markdown report with specific savings estimates
python3 {baseDir}/scripts/audit.py Options: python3 {baseDir}/scripts/audit.py --format markdown # Full report (default) python3 {baseDir}/scripts/audit.py --format summary # Quick summary only python3 {baseDir}/scripts/audit.py --dry-run # Show what would be analyzed python3 {baseDir}/scripts/audit.py --output /path/to/report.md # Save to file
Read OpenClaw config (~/.openclaw/openclaw.json or similar) List all cron jobs and their configurations List all agents and their default models Detect provider (Anthropic, OpenAI, Google, xAI) from model names
Pull cron job run history (last 7 days by default) Calculate per-job: avg tokens, avg runtime, success rate, model used Pull session history where available Calculate total token spend by model tier
Classify each task into complexity tiers: TierExamplesRecommended ModelsSimpleHealth checks, status reports, reminders, notificationsCheapest tier (Haiku, GPT-4o-mini, Flash, Grok-mini)MediumContent drafts, research, summarization, data analysisMid tier (Sonnet, GPT-4o, Pro, Grok)ComplexCoding, architecture, security review, nuanced writingTop tier (Opus, GPT-4.5, Ultra, Grok-2) Classification signals: Simple: Short output (<500 tokens), low thinking requirement, repetitive pattern, status/health tasks Medium: Medium output, some reasoning needed, creative but templated, research tasks Complex: Long output, multi-step reasoning, code generation, security-critical, tasks that previously failed on weaker models
For each task where the model tier doesn't match complexity: โ ๏ธ RECOMMENDATION: Downgrade "Knox Bot Health Check" from opus to haiku Current: anthropic/claude-opus-4 ($15/M input, $75/M output) Suggested: anthropic/claude-haiku ($0.25/M input, $1.25/M output) Reason: Simple status check averaging 300 output tokens Estimated savings: $X.XX/month Risk: LOW โ task is simple pattern matching Confidence: HIGH
Coding/development tasks Security reviews or audits Tasks that have previously failed on weaker models Tasks where the user explicitly chose a higher model Complex multi-step reasoning tasks Anything the user flagged as critical
Output a clean markdown report with: Overview โ total agents, crons, monthly spend estimate Per-agent breakdown โ model, usage, cost Per-cron breakdown โ model, frequency, avg tokens, cost Recommendations โ sorted by savings potential Total potential savings โ monthly estimate One-liner config changes โ exact model strings to swap
See references/model-pricing.md for current pricing across all providers. Update this file when prices change.
See references/task-classification.md for detailed heuristics on how tasks are classified into complexity tiers.
This skill is read-only โ it never changes your config automatically All recommendations include risk levels and confidence scores When unsure about a task's complexity, it defaults to keeping the current model The audit should be re-run periodically (monthly) as usage patterns change Token counts are estimates based on cron history โ actual costs depend on your provider's billing
Identity, auth, scanning, governance, audit, and operational guardrails.
Largest current source with strong distribution and engagement signals.