← All skills
Tencent SkillHub Β· AI

Ralph Loops

Runs autonomous iterative AI loops for requirements, planning, or building phases using structured prompts and fresh context per iteration.

skill openclawclawhub Free
0 Downloads
0 Stars
0 Installs
0 Score
High Signal

Runs autonomous iterative AI loops for requirements, planning, or building phases using structured prompts and fresh context per iteration.

⬇ 0 downloads β˜… 0 stars Unverified but indexed

Install for OpenClaw

Quick setup
  1. Download the package from Yavira.
  2. Extract the archive and review SKILL.md first.
  3. Import or place the package into your OpenClaw setup.

Requirements

Target platform
OpenClaw
Install method
Manual import
Extraction
Extract archive
Prerequisites
OpenClaw
Primary doc
SKILL.md

Package facts

Download mode
Yavira redirect
Package format
ZIP package
Source platform
Tencent SkillHub
What's included
SETUP.md, SKILL.md, dashboard/lib/cost-calculator.mjs, dashboard/lib/loop-detector.mjs, dashboard/lib/ralph-reader.mjs, dashboard/lib/transcript-reader.mjs

Validation

  • Use the Yavira download entry.
  • Review SKILL.md after the package is downloaded.
  • Confirm the extracted package contains the expected setup assets.

Install with your agent

Agent handoff

Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.

  1. Download the package from Yavira.
  2. Extract it into a folder your agent can access.
  3. Paste one of the prompts below and point your agent at the extracted folder.
New install

I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Tell me what you changed and call out any manual steps you could not complete.

Upgrade existing

I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Summarize what changed and any follow-up checks I should run.

Trust & source

Release facts

Source
Tencent SkillHub
Verification
Indexed source record
Version
1.0.2

Documentation

ClawHub primary doc Primary doc: SKILL.md 45 sections Open source page

Ralph Loops Skill

First time? Read SETUP.md first to install dependencies and verify your setup. Autonomous AI agent loops for iterative development. Based on Geoffrey Huntley's Ralph Wiggum technique, as documented by Clayton Farr. Script: skills/ralph-loops/scripts/ralph-loop.mjs Dashboard: skills/ralph-loops/dashboard/ (run with node server.mjs) Templates: skills/ralph-loops/templates/ Archive: ~/clawd/logs/ralph-archive/

Claude Code Version Compatibility

Claude Code 2.1.29 has a critical bug that spawns orphaned sub-agents consuming 99% CPU. Iterations fail with "exit code null" on first run. Fix: Downgrade to 2.1.25: npm install -g @anthropic-ai/claude-code@2.1.25 Verify: claude --version # Should show 2.1.25 This was discovered 2026-02-01. Check if newer versions fix the issue before upgrading.

⚠️ Don't Block the Conversation!

When running a Ralph loop, don't monitor it synchronously. The loop runs as a separate Claude CLI process β€” you can keep chatting. ❌ Wrong (blocks conversation): Start loop β†’ sleep 60 β†’ poll β†’ sleep 60 β†’ poll β†’ ... (6 minutes of silence) βœ… Right (stays responsive): Start loop β†’ "It's running, I'll check periodically" β†’ keep chatting β†’ check on heartbeats How to monitor without blocking: Start the loop with node ralph-loop.mjs ... (runs in background) Tell human: "Loop running. I'll check progress periodically or you can ask." Check via process poll <sessionId> when asked or during heartbeats Use the dashboard at http://localhost:3939 for real-time visibility The loop is autonomous β€” that's the whole point. Don't babysit it at the cost of ignoring your human.

Trigger Phrases

When human says: PhraseAction"Interview me about system X"Start Phase 1 requirements interview"Start planning system X"Run ./loop.sh plan (needs specs first)"Start building system X"Run ./loop.sh build (needs plan first)"Ralph loop over X"ASK which phase (see below)

When Human Says "Ralph Loop" β€” Clarify the Phase!

Don't assume which phase. Ask: "Which type of Ralph loop are we doing? 1️⃣ Interview β€” I'll ask you questions to build specs (Phase 1) 2️⃣ Planning β€” I'll iterate on an implementation plan (Phase 2) 3️⃣ Building β€” I'll implement from a plan, one task per iteration (Phase 3) 4️⃣ Generic β€” Simple iterative refinement on a single topic" Then proceed based on their answer: ChoiceActionInterviewUse templates/requirements-interview.md protocolPlanningNeed specs first β†’ run planning loop with PROMPT_plan.mdBuildingNeed plan first β†’ run build loop with PROMPT_build.mdGenericCreate prompt file, run ralph-loop.mjs directly

Generic Ralph Loop Flow (Phase 4)

For simple iterative refinement (not full system builds): Clarify the task β€” What exactly should be improved/refined? Create a prompt file β€” Save to /tmp/ralph-prompt-<task>.md Set completion criteria β€” What signals "done"? Run the loop: node skills/ralph-loops/scripts/ralph-loop.mjs \ --prompt "/tmp/ralph-prompt-<task>.md" \ --model opus \ --max 10 \ --done "RALPH_DONE" Or spawn as sub-agent for long-running tasks

Core Philosophy

"Human roles shift from 'telling the agent what to do' to 'engineering conditions where good outcomes emerge naturally through iteration." β€” Clayton Farr Three principles drive everything: Context is scarce β€” With ~176K usable tokens from a 200K window, keep each iteration lean Plans are disposable β€” A drifting plan is cheaper to regenerate than salvage Backpressure beats direction β€” Engineer environments where wrong outputs get rejected automatically

Three-Phase Workflow

β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β” β”‚ Phase 1: REQUIREMENTS β”‚ β”‚ Human + LLM conversation β†’ JTBD β†’ Topics β†’ specs/*.md β”‚ β”œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€ β”‚ Phase 2: PLANNING β”‚ β”‚ Gap analysis (specs vs code) β†’ IMPLEMENTATION_PLAN.md β”‚ β”œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€ β”‚ Phase 3: BUILDING β”‚ β”‚ One task per iteration β†’ fresh context β†’ backpressure β†’ commit β”‚ β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜

Phase 1: Requirements (Talk to Human)

Goal: Understand what to build BEFORE building it. This is the most important phase. Use structured conversation to: Identify Jobs to Be Done (JTBD) What user need or outcome are we solving? Not features β€” outcomes Break JTBD into Topics of Concern Each topic = one distinct aspect/component Use the "one sentence without 'and'" test βœ“ "The color extraction system analyzes images to identify dominant colors" βœ— "The user system handles authentication, profiles, and billing" β†’ 3 topics Create Specs for Each Topic One markdown file per topic in specs/ Capture requirements, acceptance criteria, edge cases Template: templates/requirements-interview.md

Phase 2: Planning (Gap Analysis)

Goal: Create a prioritized task list without implementing anything. Uses PROMPT_plan.md in the loop: Study all specs Study existing codebase Compare specs vs code (gap analysis) Generate IMPLEMENTATION_PLAN.md with prioritized tasks NO implementation β€” planning only Usually completes in 1-2 iterations.

Phase 3: Building (One Task Per Iteration)

Goal: Implement tasks one at a time with fresh context. Uses PROMPT_build.md in the loop: Read IMPLEMENTATION_PLAN.md Pick the most important task Investigate codebase (don't assume not implemented) Implement Run validation (backpressure) Update plan, commit Exit β†’ fresh context β†’ next iteration Key insight: One task per iteration keeps context lean. The agent stays in the "smart zone" instead of accumulating cruft. Why fresh context matters: No accumulated mistakes β€” Each iteration starts clean; previous errors don't compound Full context budget β€” 200K tokens for THIS task, not shared with finished work Reduced hallucination β€” Shorter contexts = more grounded responses Natural checkpoints β€” Each commit is a save point; easy to revert single iterations

File Structure

project/ β”œβ”€β”€ loop.sh # Ralph loop script β”œβ”€β”€ PROMPT_plan.md # Planning mode instructions β”œβ”€β”€ PROMPT_build.md # Building mode instructions β”œβ”€β”€ AGENTS.md # Operational guide (~60 lines max) β”œβ”€β”€ IMPLEMENTATION_PLAN.md # Prioritized task list (generated) └── specs/ # Requirement specs β”œβ”€β”€ topic-a.md β”œβ”€β”€ topic-b.md └── ...

File Purposes

FilePurposeWho Createsspecs/*.mdSource of truth for requirementsHuman + Phase 1PROMPT_plan.mdInstructions for planning modeCopy from templatePROMPT_build.mdInstructions for building modeCopy from templateAGENTS.mdBuild/test/lint commandsHuman + RalphIMPLEMENTATION_PLAN.mdTask list with prioritiesRalph (Phase 2)

Project Organization (Systems)

For Clawdbot systems, each Ralph project lives in <workspace>/systems/<name>/: systems/ β”œβ”€β”€ health-tracker/ # Example system β”‚ β”œβ”€β”€ specs/ β”‚ β”‚ β”œβ”€β”€ daily-tracking.md β”‚ β”‚ └── test-scheduling.md β”‚ β”œβ”€β”€ PROMPT_plan.md β”‚ β”œβ”€β”€ PROMPT_build.md β”‚ β”œβ”€β”€ AGENTS.md β”‚ β”œβ”€β”€ IMPLEMENTATION_PLAN.md # ← exists = past Phase 1 β”‚ └── src/ └── activity-planner/ β”œβ”€β”€ specs/ # ← empty = still in Phase 1 └── ...

Phase Detection (Auto)

Detect current phase by checking what files exist: What ExistsCurrent PhaseNext ActionNothing / empty specs/Phase 1: RequirementsRun requirements interviewspecs/*.md but no IMPLEMENTATION_PLAN.mdReady for Phase 2Run ./loop.sh planspecs/*.md + IMPLEMENTATION_PLAN.mdPhase 2 or 3Review plan, run ./loop.sh buildPlan shows all tasks completeDoneArchive or iterate Quick check: # What phase are we in? [ -z "$(ls specs/ 2>/dev/null)" ] && echo "Phase 1: Need specs" && exit [ ! -f IMPLEMENTATION_PLAN.md ] && echo "Phase 2: Need plan" && exit echo "Phase 3: Ready to build (or done)"

JTBD Breakdown

The hierarchy matters: JTBD (Job to Be Done) └── Topic of Concern (1 per spec file) └── Tasks (many per topic, in IMPLEMENTATION_PLAN.md) Example: JTBD: "Help designers create mood boards" Topics: Image collection β†’ specs/image-collection.md Color extraction β†’ specs/color-extraction.md Layout system β†’ specs/layout-system.md Sharing β†’ specs/sharing.md Tasks: Each spec generates multiple implementation tasks

Topic Scope Test

Can you describe the topic in one sentence without "and"? If you need "and" or "also", it's probably multiple topics. Split it. When to split: Multiple verbs in the description β†’ separate topics Different user personas involved β†’ separate topics Could be implemented by different teams β†’ separate topics Has its own failure modes β†’ probably its own topic Example split: ❌ "User management handles registration, authentication, profiles, and permissions" βœ… Split into: - "Registration creates new user accounts from email/password" - "Authentication verifies user identity via login flow" - "Profiles let users view and edit their information" - "Permissions control what actions users can perform" Counter-example (don't split): βœ… Keep together: "Color extraction analyzes images and returns dominant color palettes" Why: "analyzes" and "returns" are steps in one operation, not separate concerns.

Backpressure Mechanisms

Autonomous loops converge when wrong outputs get rejected. Three layers:

1. Downstream Gates (Hard)

  • Tests, type-checking, linting, build validation. Deterministic.
  • # In AGENTS.md
  • ## Validation
  • Tests: `npm test`
  • Typecheck: `npm run typecheck`
  • Lint: `npm run lint`

2. Upstream Steering (Soft)

Existing code patterns guide the agent. It discovers conventions through exploration.

3. LLM-as-Judge (Subjective)

For subjective criteria (tone, UX, aesthetics), use another LLM call with binary pass/fail. Start with hard gates. Add LLM-as-judge for subjective criteria only after mechanical backpressure works.

Prompt Structure

Geoffrey's prompts follow a numbered pattern: SectionPurpose0a-0dOrient: Study specs, source, current plan1-4Main instructions: What to do this iteration999+Guardrails: Invariants (higher number = more critical)

The Numbered Guardrails Pattern

Guardrails use escalating numbers (99999, 999999, 9999999...) to signal priority: 99999. Important: Capture the why in documentation. 999999. Important: Single sources of truth, no migrations. 9999999. Create git tags after successful builds. 99999999. Add logging if needed to debug. 999999999. Keep IMPLEMENTATION_PLAN.md current. Why this works: Visual prominence β€” Large numbers stand out, harder to skip Implicit priority β€” More 9s = more critical (like DEFCON levels in reverse) No collisions β€” Sparse numbering lets you insert new rules without renumbering Mnemonic β€” Claude treats these as invariants, not suggestions The "Important:" prefix is deliberate β€” it triggers Claude's attention.

Key Language Patterns

Use Geoffrey's specific phrasing β€” it matters: "study" (not "read" or "look at") "don't assume not implemented" (critical!) "using parallel subagents" / "up to N subagents" "only 1 subagent for build/tests" (backpressure control) "Ultrathink" (deep reasoning trigger) "capture the why" "keep it up to date" "resolve them or document them"

1. Set Up Project Structure

mkdir -p myproject/specs cd myproject git init # Ralph expects git for commits # Copy templates cp .//templates/PROMPT_plan.md . cp .//templates/PROMPT_build.md . cp .//templates/AGENTS.md . cp .//templates/loop.sh . chmod +x loop.sh

2. Customize Templates (Required!)

PROMPT_plan.md β€” Replace [PROJECT_GOAL] with your actual goal: # Before: ULTIMATE GOAL: We want to achieve [PROJECT_GOAL]. # After: ULTIMATE GOAL: We want to achieve a fully functional mood board app with image upload and color extraction. PROMPT_build.md β€” Adjust source paths if not using src/: # Before: 0c. For reference, the application source code is in `src/*`. # After: 0c. For reference, the application source code is in `lib/*`. AGENTS.md β€” Update build/test/lint commands for your stack.

3. Phase 1: Requirements Gathering (Don't Skip!)

This phase happens WITH the human. Use the interview template: cat .//templates/requirements-interview.md The workflow: Discuss the JTBD (Job to Be Done) β€” outcomes, not features Break into Topics of Concern (each passes the "one sentence" test) Write a spec file for each topic: specs/topic-name.md Human reviews and approves specs Example output: specs/ β”œβ”€β”€ image-collection.md β”œβ”€β”€ color-extraction.md β”œβ”€β”€ layout-system.md └── sharing.md

4. Phase 2: Planning

./loop.sh plan Wait for IMPLEMENTATION_PLAN.md to be generated (usually 1-2 iterations). Review it β€” this is your task list.

5. Phase 3: Building

./loop.sh build 20 # Max 20 iterations Watch it work. Add backpressure (tests, lints) as patterns emerge. Check commits for progress.

Loop Script Options

./loop.sh # Build mode, unlimited ./loop.sh 20 # Build mode, max 20 iterations ./loop.sh plan # Plan mode, unlimited ./loop.sh plan 5 # Plan mode, max 5 iterations Or use the Node.js wrapper for more control: node skills/ralph-loops/scripts/ralph-loop.mjs \ --prompt "./PROMPT_build.md" \ --model opus \ --max 20 \ --done "RALPH_DONE"

When to Regenerate the Plan

Plans drift. Regenerate when: Ralph is going off track (implementing wrong things) Plan feels stale or doesn't match current state Too much clutter from completed items You've made significant spec changes You're confused about what's actually done Just switch back to planning mode: ./loop.sh plan Regeneration cost is one Planning loop. Cheap compared to Ralph going in circles.

Safety

Ralph requires --dangerously-skip-permissions to run autonomously. This bypasses Claude's permission system entirely. Philosophy: "It's not if it gets popped, it's when. And what is the blast radius?" Protections: Run in isolated environments (Docker, VM) Only the API keys needed for the task No access to private data beyond requirements Restrict network connectivity where possible Escape hatches: Ctrl+C stops the loop; git reset --hard reverts uncommitted changes

Cost Expectations

Task TypeModelIterationsEst. CostGenerate planOpus1-2$0.50-1.00Implement simple featureOpus3-5$1.00-2.00Implement complex featureOpus10-20$3.00-8.00Full project buildoutOpus50+$15-50+ Tip: Use Sonnet for simpler tasks where plan is clear. Use Opus for planning and complex reasoning.

Real-World Results

From Geoffrey Huntley: 6 repos generated overnight at YC hackathon $50k contract completed for $297 in API costs Created entire programming language over 3 months

Advanced: Running as Sub-Agent

For long loops, spawn as sub-agent so main session stays responsive: sessions_spawn({ task: `cd /path/to/project && ./loop.sh build 20 Summarize what was implemented when done.`, label: "ralph-build", model: "opus" }) Check progress: sessions_list({ kinds: ["spawn"] }) sessions_history({ label: "ralph-build", limit: 5 })

Ralph keeps implementing the same thing

Plan is stale β†’ regenerate with ./loop.sh plan Backpressure missing β†’ add tests that catch duplicates

Ralph goes in circles

Add more specific guardrails to prompts Check if specs are ambiguous Regenerate plan

Context getting bloated

Ensure one task per iteration (check prompt) Keep AGENTS.md under 60 lines Move status/progress to IMPLEMENTATION_PLAN.md, not AGENTS.md

Tests not running

Check AGENTS.md has correct validation commands Ensure backpressure section in prompt references AGENTS.md

Projects Without Git

The loop script expects git for commits and pushes. For projects without version control: Option 1: Initialize git anyway (recommended) git init git add -A git commit -m "Initial commit before Ralph" Option 2: Modify the prompts Remove git-related guardrails from PROMPT_build.md Remove the git push section from loop.sh Use file backups instead: add cp -r src/ backups/iteration-$ITERATION/ to loop.sh Option 3: Use tarball snapshots # Add to loop.sh before each iteration: tar -czf "snapshots/pre-iteration-$ITERATION.tar.gz" src/

Very Large Codebases

For codebases with 100K+ lines: Reduce subagent parallelism: Change "up to 500 parallel Sonnet subagents" to "up to 50" in prompts Scope narrowly: Use focused specs that target specific directories Add path restrictions: In AGENTS.md, note which directories are in-scope Consider workspace splitting: Treat large modules as separate Ralph projects

When Claude CLI Isn't Available

The methodology works with any Claude interface: Claude API directly: # Replace loop.sh with API calls using curl or a script curl https://api.anthropic.com/v1/messages \ -H "x-api-key: $ANTHROPIC_API_KEY" \ -H "content-type: application/json" \ -d '{"model": "claude-sonnet-4-20250514", "max_tokens": 8192, "messages": [...]}' Alternative agents: Aider: aider --opus --auto-commits Continue.dev: Use with Claude API key Cursor: Composer mode with PROMPT files as context The key principles (one task per iteration, fresh context, backpressure) apply regardless of tooling.

Non-Node.js Projects

Adapt AGENTS.md for your stack: StackBuildTestLintPythonpip install -e .pytestruff .Gogo build ./...go test ./...golangci-lint runRustcargo buildcargo testcargo clippyRubybundle installrspecrubocop Also update path references in prompts (src/* β†’ your source directory).

Learn More

Geoffrey Huntley: https://ghuntley.com/ralph/ Clayton Farr's Playbook: https://github.com/ClaytonFarr/ralph-playbook Geoffrey's Fork: https://github.com/ghuntley/how-to-ralph-wiggum

Credits

Built by Johnathan & Q β€” a human-AI dyad. Twitter: @spacepixel ClawdHub: clawhub.ai/skills/ralph-loops

Category context

Agent frameworks, memory systems, reasoning layers, and model-native orchestration.

Source: Tencent SkillHub

Largest current source with strong distribution and engagement signals.

Package contents

Included in package
4 Scripts2 Docs
  • SKILL.md Primary doc
  • SETUP.md Docs
  • dashboard/lib/cost-calculator.mjs Scripts
  • dashboard/lib/loop-detector.mjs Scripts
  • dashboard/lib/ralph-reader.mjs Scripts
  • dashboard/lib/transcript-reader.mjs Scripts