Requirements
- Target platform
- OpenClaw
- Install method
- Manual import
- Extraction
- Extract archive
- Prerequisites
- OpenClaw
- Primary doc
- SKILL.md
Comprehensive deep reasoning framework that guides systematic, thorough thinking for complex tasks. Automatically applies for multi-step problems, ambiguous...
Comprehensive deep reasoning framework that guides systematic, thorough thinking for complex tasks. Automatically applies for multi-step problems, ambiguous...
Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.
I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Tell me what you changed and call out any manual steps you could not complete.
I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Summarize what changed and any follow-up checks I should run.
Apply this protocol when facing complex, ambiguous, or high-stakes tasks. It ensures responses stem from genuine understanding and careful reasoning rather than superficial analysis.
Activate this protocol when: The task has multiple valid approaches with meaningful trade-offs Requirements are ambiguous or underspecified The problem involves architectural or design decisions Debugging requires systematic investigation The task touches multiple systems or files Stakes are high (data integrity, security, production impact) The user explicitly asks to think carefully or deeply Skip for trivial, single-step tasks with obvious solutions.
Your reasoning should be organic and exploratory, not mechanical: Think like a detective following leads, not a robot following steps Let each realization lead naturally to the next Show genuine curiosity β "Wait, what if...", "Actually, this changes things..." Avoid formulaic analysis; adapt your thinking style to the problem Errors in reasoning are opportunities for deeper understanding, not just corrections to make Never feel forced or structured β the steps below are a guide, not a rigid sequence
Scale analysis depth based on: Query complexity: Simple lookup vs. multi-dimensional problem Stakes involved: Low-risk formatting vs. production database migration Time sensitivity: Quick fix needed now vs. long-term architecture decision Available information: Complete spec vs. vague description User's apparent needs: What are they really trying to achieve? Adjust thinking style based on: Technical vs. conceptual: Implementation detail vs. architecture decision Analytical vs. exploratory: Clear bug with stack trace vs. vague performance issue Abstract vs. concrete: Design pattern selection vs. specific function implementation Single vs. multi-scope: One file change vs. cross-module refactor
Rephrase the problem in your own words to verify understanding Identify what is known vs. unknown Consider the broader context β why is this question being asked? What's the underlying goal? Map out what knowledge or codebase areas are needed to address this Flag ambiguities that need clarification before proceeding
Break the task into core components Identify explicit and implicit requirements Map constraints and limitations Define what a successful outcome looks like
Generate at least 2-3 possible approaches before committing Keep multiple working hypotheses active β don't collapse to one prematurely Consider unconventional or non-obvious interpretations Look for creative combinations of different approaches Evaluate trade-offs: complexity, performance, maintainability, risk Show why certain approaches are more suitable than others
Think like a detective β each realization should lead naturally to the next: Start with obvious aspects, then dig deeper Notice patterns and connections across the codebase Question initial assumptions as understanding develops Circle back to earlier ideas with new context Build progressively deeper insights Be open to serendipitous insights β unexpected connections often reveal the best solutions Follow interesting tangents, but tie them back to the core issue
Test conclusions against evidence (code, docs, tests) Look for edge cases and potential failure modes Actively seek counter-examples that could disprove your current theory When finding mistakes in reasoning, acknowledge naturally and show how new understanding develops β view errors as opportunities for deeper insight Cross-check for logical consistency Verify completeness: "Have I addressed the full scope?"
Connect findings into a coherent picture Identify key principles or patterns that emerged Create useful abstractions β turn findings into reusable concepts or guidelines Note important implications and downstream effects Ensure the synthesis answers the original question
Apply the same careful analysis at both macro (system/architecture) and micro (function/logic) levels Use patterns recognized at one scale to inform analysis at another Maintain consistency while allowing for scale-appropriate methods Show how detailed analysis supports or challenges broader conclusions
While exploring related ideas: Maintain clear connection to the original query at all times When following tangents, explicitly tie them back to the core issue Periodically ask: "Is this exploration serving the final response?" Keep sight of the user's actual goal, not just the literal question Ensure all exploration serves the final response
Before delivering a response, verify: All aspects of the original question are addressed Conclusions are supported by evidence (not assumptions) Edge cases and failure modes are considered Trade-offs are explicitly stated The recommended approach is justified over alternatives No logical inconsistencies in the reasoning Detail level matches the user's apparent expertise and needs Likely follow-up questions are anticipated
Anti-PatternInstead DoJumping to implementation immediatelyAnalyze the problem space firstConsidering only one approachGenerate and compare alternativesIgnoring edge casesActively seek boundary conditionsAssuming without verifyingRead the code, check the docsOver-engineering simple tasksMatch depth to complexityAnalysis paralysis on trivial decisionsSet a time-box, then decideDrawing premature conclusionsVerify with evidence before committingNot seeking counter-examplesActively look for cases that disprove your theoryMechanical checklist thinkingLet reasoning flow organically; adapt to the problem
Evaluate your thinking against: Completeness: Did I cover all dimensions of the problem? Logical consistency: Do my conclusions follow from my analysis? Evidence support: Are claims backed by code, docs, or reasoning? Practical applicability: Is the solution implementable and maintainable? Clarity: Can the reasoning be followed and verified?
During extended analysis, maintain awareness of: What has been established so far What remains to be determined Current confidence level in conclusions Open questions or uncertainties Whether the current approach is productive or needs pivoting
For detailed examples of thinking patterns, natural language flow, and domain-specific applications, see reference.md.
Agent frameworks, memory systems, reasoning layers, and model-native orchestration.
Largest current source with strong distribution and engagement signals.