← All skills
Tencent SkillHub · AI

claw-compactor

Claw Compactor v6.0 — 50%+ savings through rule-based compression, dictionary encoding, session observation compression, and progressive context loading.

skill openclawclawhub Free
0 Downloads
0 Stars
0 Installs
0 Score
High Signal

Claw Compactor v6.0 — 50%+ savings through rule-based compression, dictionary encoding, session observation compression, and progressive context loading.

⬇ 0 downloads ★ 0 stars Unverified but indexed

Install for OpenClaw

Quick setup
  1. Download the package from Yavira.
  2. Extract the archive and review SKILL.md first.
  3. Import or place the package into your OpenClaw setup.

Requirements

Target platform
OpenClaw
Install method
Manual import
Extraction
Extract archive
Prerequisites
OpenClaw
Primary doc
SKILL.md

Package facts

Download mode
Yavira redirect
Package format
ZIP package
Source platform
Tencent SkillHub
What's included
pyproject.toml, README.md, SKILL.md, scripts/mem_compress.py, scripts/estimate_tokens.py, scripts/dedup_memory.py

Validation

  • Use the Yavira download entry.
  • Review SKILL.md after the package is downloaded.
  • Confirm the extracted package contains the expected setup assets.

Install with your agent

Agent handoff

Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.

  1. Download the package from Yavira.
  2. Extract it into a folder your agent can access.
  3. Paste one of the prompts below and point your agent at the extracted folder.
New install

I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Then review README.md for any prerequisites, environment setup, or post-install checks. Tell me what you changed and call out any manual steps you could not complete.

Upgrade existing

I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Then review README.md for any prerequisites, environment setup, or post-install checks. Summarize what changed and any follow-up checks I should run.

Trust & source

Release facts

Source
Tencent SkillHub
Verification
Indexed source record
Version
6.0.0

Documentation

ClawHub primary doc Primary doc: SKILL.md 16 sections Open source page

🦞 Claw Compactor

"Cut your tokens. Keep your facts." Cut your AI agent's token spend in half. One command compresses your entire workspace — memory files, session transcripts, sub-agent context — using 5 layered compression techniques. Deterministic. Mostly lossless. No LLM required.

Features

5 compression layers working in sequence for maximum savings Zero LLM cost — all compression is rule-based and deterministic Lossless roundtrip for dictionary, RLE, and rule-based compression ~97% savings on session transcripts via observation extraction Tiered summaries (L0/L1/L2) for progressive context loading CJK-aware — full Chinese/Japanese/Korean support One command (full) runs everything in optimal order

5 Compression Layers

#LayerMethodSavingsLossless?1Rule engineDedup lines, strip markdown filler, merge sections4-8%✅2Dictionary encodingAuto-learned codebook, $XX substitution4-5%✅3Observation compressionSession JSONL → structured summaries~97%❌*4RLE patternsPath shorthand ($WS), IP prefix, enum compaction1-2%✅5Compressed Context Protocolultra/medium/light abbreviation20-60%❌* *Lossy techniques preserve all facts and decisions; only verbose formatting is removed.

Quick Start

git clone https://github.com/aeromomo/claw-compactor.git cd claw-compactor # See how much you'd save (non-destructive) python3 scripts/mem_compress.py /path/to/workspace benchmark # Compress everything python3 scripts/mem_compress.py /path/to/workspace full Requirements: Python 3.9+. Optional: pip install tiktoken for exact token counts (falls back to heuristic).

Architecture

┌─────────────────────────────────────────────────────────────┐ │ mem_compress.py │ │ (unified entry point) │ └──────┬──────┬──────┬──────┬──────┬──────┬──────┬──────┬────┘ │ │ │ │ │ │ │ │ ▼ ▼ ▼ ▼ ▼ ▼ ▼ ▼ estimate compress dict dedup observe tiers audit optimize └──────┴──────┴──┬───┴──────┴──────┴──────┴──────┘ ▼ ┌────────────────┐ │ lib/ │ │ tokens.py │ ← tiktoken or heuristic │ markdown.py │ ← section parsing │ dedup.py │ ← shingle hashing │ dictionary.py │ ← codebook compression │ rle.py │ ← path/IP/enum encoding │ tokenizer_ │ │ optimizer.py │ ← format optimization │ config.py │ ← JSON config │ exceptions.py │ ← error types └────────────────┘

Commands

All commands: python3 scripts/mem_compress.py <workspace> <command> [options] CommandDescriptionTypical SavingsfullComplete pipeline (all steps in order)50%+ combinedbenchmarkDry-run performance report—compressRule-based compression4-8%dictDictionary encoding with auto-codebook4-5%observeSession transcript → observations~97%tiersGenerate L0/L1/L2 summaries88-95% on sub-agent loadsdedupCross-file duplicate detectionvariesestimateToken count report—auditWorkspace health check—optimizeTokenizer-level format fixes1-3%

Global Options

--json — Machine-readable JSON output --dry-run — Preview changes without writing --since YYYY-MM-DD — Filter sessions by date --auto-merge — Auto-merge duplicates (dedup)

Real-World Savings

Workspace StateTypical SavingsNotesSession transcripts (observe)~97%Megabytes of JSONL → concise observation MDVerbose/new workspace50-70%First run on unoptimized workspaceRegular maintenance10-20%Weekly runs on active workspaceAlready-optimized3-12%Diminishing returns — workspace is clean

cacheRetention — Complementary Optimization

Before compression runs, enable prompt caching for a 90% discount on cached tokens: { "models": { "model-name": { "cacheRetention": "long" } } } Compression reduces token count, caching reduces cost-per-token. Together: 50% compression + 90% cache discount = 95% effective cost reduction.

Heartbeat Automation

  • Run weekly or on heartbeat:
  • ## Memory Maintenance (weekly)
  • python3 skills/claw-compactor/scripts/mem_compress.py <workspace> benchmark
  • If savings > 5%: run full pipeline
  • If pending transcripts: run observe
  • Cron example:
  • 0 3 * * 0 cd /path/to/skills/claw-compactor && python3 scripts/mem_compress.py /path/to/workspace full

Configuration

Optional claw-compactor-config.json in workspace root: { "chars_per_token": 4, "level0_max_tokens": 200, "level1_max_tokens": 500, "dedup_similarity_threshold": 0.6, "dedup_shingle_size": 3 } All fields optional — sensible defaults are used when absent.

Artifacts

FilePurposememory/.codebook.jsonDictionary codebook (must travel with memory files)memory/.observed-sessions.jsonTracks processed transcriptsmemory/observations/Compressed session summariesmemory/MEMORY-L0.mdLevel 0 summary (~200 tokens)

FAQ

Q: Will compression lose my data? A: Rule engine, dictionary, RLE, and tokenizer optimization are fully lossless. Observation compression and CCP are lossy but preserve all facts and decisions. Q: How does dictionary decompression work? A: decompress_text(text, codebook) expands all $XX codes back. The codebook JSON must be present. Q: Can I run individual steps? A: Yes. Every command is independent: compress, dict, observe, tiers, dedup, optimize. Q: What if tiktoken isn't installed? A: Falls back to a CJK-aware heuristic (chars÷4). Results are ~90% accurate. Q: Does it handle Chinese/Japanese/Unicode? A: Yes. Full CJK support including character-aware token estimation and Chinese punctuation normalization.

Troubleshooting

FileNotFoundError on workspace: Ensure path points to workspace root (contains memory/ or MEMORY.md) Dictionary decompression fails: Check memory/.codebook.json exists and is valid JSON Zero savings on benchmark: Workspace is already optimized — nothing to do observe finds no transcripts: Check sessions directory for .jsonl files Token count seems wrong: Install tiktoken: pip3 install tiktoken

Credits

Inspired by claude-mem by thedotmack Built by Bot777 🤖 for OpenClaw

License

MIT

Category context

Agent frameworks, memory systems, reasoning layers, and model-native orchestration.

Source: Tencent SkillHub

Largest current source with strong distribution and engagement signals.

Package contents

Included in package
3 Scripts2 Docs1 Files
  • SKILL.md Primary doc
  • README.md Docs
  • scripts/dedup_memory.py Scripts
  • scripts/estimate_tokens.py Scripts
  • scripts/mem_compress.py Scripts
  • pyproject.toml Files