Requirements
- Target platform
- OpenClaw
- Install method
- Manual import
- Extraction
- Extract archive
- Prerequisites
- OpenClaw
- Primary doc
- SKILL.md
Routes LLM requests to the cheapest capable model across 8 providers (Anthropic, Google, OpenAI, DeepSeek, xAI, Moonshot, Mistral, Ollama) and 25+ models. Scores prompts on 8 dimensions in under 1ms, supports three routing modes (eco, standard, gigachad), and logs all decisions for cost tracking.
Routes LLM requests to the cheapest capable model across 8 providers (Anthropic, Google, OpenAI, DeepSeek, xAI, Moonshot, Mistral, Ollama) and 25+ models. Scores prompts on 8 dimensions in under 1ms, supports three routing modes (eco, standard, gigachad), and logs all decisions for cost tracking.
Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.
I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Then review README.md for any prerequisites, environment setup, or post-install checks. Tell me what you changed and call out any manual steps you could not complete.
I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Then review README.md for any prerequisites, environment setup, or post-install checks. Summarize what changed and any follow-up checks I should run.
Route every LLM request to the cheapest model that can handle it. Stop paying Opus prices for "hello" and "summarize this." Supports 8 providers and 25+ models: Anthropic (Claude), Google (Gemini), OpenAI (GPT / o-series), xAI (Grok), DeepSeek, Moonshot (Kimi), Mistral, and Ollama (local).
Your prompt arrives The classifier scores it on 8 dimensions (token count, code presence, reasoning markers, simplicity indicators, multi-step patterns, question count, system prompt complexity, conversation depth) in under 1 millisecond The router maps the resulting tier (simple / standard / complex) to a model based on your active mode and configured providers The request is proxied to the correct API The routing decision and cost are logged to a local JSONL file
ModeSimpleStandardComplexecoGrok 4.1 FastGemini FlashHaikustandardGrok 4.1 FastHaikuSonnetgigachadHaikuSonnetOpus 4.6 Each cell shows the first-choice model. The router tries a preference list and falls through to the next available provider if the first is not configured.
CommandWhat It Doesroute_requestSend a prompt and get a response from the cheapest capable modelclassify_promptAnalyze prompt complexity without making an LLM callget_routing_statsView cost savings and model distribution statsget_configView current configuration (keys redacted)set_modeChange routing mode at runtimeget_recent_routing_logInspect recent routing decisions
Heartbeats and summaries always route to the cheapest model Type /opus, /sonnet, /haiku, /flash, or /grok-fast to force a specific model Sub-agent calls automatically step down one tier from their parent
Get at least one API key (Anthropic or Google required; others optional): Anthropic: https://console.anthropic.com/settings/keys Google AI: https://aistudio.google.com/app/apikey xAI: https://console.x.ai OpenAI: https://platform.openai.com/api-keys DeepSeek: https://platform.deepseek.com Moonshot: https://platform.moonshot.cn Mistral: https://console.mistral.ai Run the setup script: npm run setup Choose your routing mode (eco / standard / gigachad)
Prompt content is never stored. Only a SHA-256 hash is logged. All data stays local in ~/.config/clawd-throttle/ API keys stored in your local config file
Agent frameworks, memory systems, reasoning layers, and model-native orchestration.
Largest current source with strong distribution and engagement signals.