Requirements
- Target platform
- OpenClaw
- Install method
- Manual import
- Extraction
- Extract archive
- Prerequisites
- OpenClaw
- Primary doc
- SKILL.md
Routes LLM requests to the cheapest capable model across 8 providers (Anthropic, Google, OpenAI, DeepSeek, xAI, Moonshot, Mistral, Ollama) and 25+ models. Scores prompts on 8 dimensions in under 1ms, supports three routing modes (eco, standard, gigachad), and logs all decisions for cost tracking.
Routes LLM requests to the cheapest capable model across 8 providers (Anthropic, Google, OpenAI, DeepSeek, xAI, Moonshot, Mistral, Ollama) and 25+ models. Scores prompts on 8 dimensions in under 1ms, supports three routing modes (eco, standard, gigachad), and logs all decisions for cost tracking.
This item is timing out or returning errors right now. Review the source page and try again later.
Use the source page and any available docs to guide the install because the item is currently unstable or timing out.
I tried to install a skill package from Yavira, but the item is currently unstable or timing out. Inspect the source page and any extracted docs, then tell me what you can confirm and any manual steps still required. Then review README.md for any prerequisites, environment setup, or post-install checks.
I tried to upgrade a skill package from Yavira, but the item is currently unstable or timing out. Compare the source page and any extracted docs with my current installation, then summarize what changed and what manual follow-up I still need. Then review README.md for any prerequisites, environment setup, or post-install checks.
Route every LLM request to the cheapest model that can handle it. Stop paying Opus prices for "hello" and "summarize this." Supports 8 providers and 25+ models: Anthropic (Claude), Google (Gemini), OpenAI (GPT / o-series), xAI (Grok), DeepSeek, Moonshot (Kimi), Mistral, and Ollama (local).
Your prompt arrives The classifier scores it on 8 dimensions (token count, code presence, reasoning markers, simplicity indicators, multi-step patterns, question count, system prompt complexity, conversation depth) in under 1 millisecond The router maps the resulting tier (simple / standard / complex) to a model based on your active mode and configured providers The request is proxied to the correct API The routing decision and cost are logged to a local JSONL file
ModeSimpleStandardComplexecoGrok 4.1 FastGemini FlashHaikustandardGrok 4.1 FastHaikuSonnetgigachadHaikuSonnetOpus 4.6 Each cell shows the first-choice model. The router tries a preference list and falls through to the next available provider if the first is not configured.
CommandWhat It Doesroute_requestSend a prompt and get a response from the cheapest capable modelclassify_promptAnalyze prompt complexity without making an LLM callget_routing_statsView cost savings and model distribution statsget_configView current configuration (keys redacted)set_modeChange routing mode at runtimeget_recent_routing_logInspect recent routing decisions
Heartbeats and summaries always route to the cheapest model Type /opus, /sonnet, /haiku, /flash, or /grok-fast to force a specific model Sub-agent calls automatically step down one tier from their parent
Get at least one API key (Anthropic or Google required; others optional): Anthropic: https://console.anthropic.com/settings/keys Google AI: https://aistudio.google.com/app/apikey xAI: https://console.x.ai OpenAI: https://platform.openai.com/api-keys DeepSeek: https://platform.deepseek.com Moonshot: https://platform.moonshot.cn Mistral: https://console.mistral.ai Run the setup script: npm run setup Choose your routing mode (eco / standard / gigachad)
Prompt content is never stored. Only a SHA-256 hash is logged. All data stays local in ~/.config/clawd-throttle/ API keys stored in your local config file
Agent frameworks, memory systems, reasoning layers, and model-native orchestration.
Largest current source with strong distribution and engagement signals.