Requirements
- Target platform
- OpenClaw
- Install method
- Manual import
- Extraction
- Extract archive
- Prerequisites
- OpenClaw
- Primary doc
- SKILL.md
Claw Compactor v6.0 — 50%+ savings through rule-based compression, dictionary encoding, session observation compression, and progressive context loading.
Claw Compactor v6.0 — 50%+ savings through rule-based compression, dictionary encoding, session observation compression, and progressive context loading.
Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.
I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Then review README.md for any prerequisites, environment setup, or post-install checks. Tell me what you changed and call out any manual steps you could not complete.
I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Then review README.md for any prerequisites, environment setup, or post-install checks. Summarize what changed and any follow-up checks I should run.
"Cut your tokens. Keep your facts." Cut your AI agent's token spend in half. One command compresses your entire workspace — memory files, session transcripts, sub-agent context — using 5 layered compression techniques. Deterministic. Mostly lossless. No LLM required.
5 compression layers working in sequence for maximum savings Zero LLM cost — all compression is rule-based and deterministic Lossless roundtrip for dictionary, RLE, and rule-based compression ~97% savings on session transcripts via observation extraction Tiered summaries (L0/L1/L2) for progressive context loading CJK-aware — full Chinese/Japanese/Korean support One command (full) runs everything in optimal order
#LayerMethodSavingsLossless?1Rule engineDedup lines, strip markdown filler, merge sections4-8%✅2Dictionary encodingAuto-learned codebook, $XX substitution4-5%✅3Observation compressionSession JSONL → structured summaries~97%❌*4RLE patternsPath shorthand ($WS), IP prefix, enum compaction1-2%✅5Compressed Context Protocolultra/medium/light abbreviation20-60%❌* *Lossy techniques preserve all facts and decisions; only verbose formatting is removed.
git clone https://github.com/aeromomo/claw-compactor.git cd claw-compactor # See how much you'd save (non-destructive) python3 scripts/mem_compress.py /path/to/workspace benchmark # Compress everything python3 scripts/mem_compress.py /path/to/workspace full Requirements: Python 3.9+. Optional: pip install tiktoken for exact token counts (falls back to heuristic).
┌─────────────────────────────────────────────────────────────┐ │ mem_compress.py │ │ (unified entry point) │ └──────┬──────┬──────┬──────┬──────┬──────┬──────┬──────┬────┘ │ │ │ │ │ │ │ │ ▼ ▼ ▼ ▼ ▼ ▼ ▼ ▼ estimate compress dict dedup observe tiers audit optimize └──────┴──────┴──┬───┴──────┴──────┴──────┴──────┘ ▼ ┌────────────────┐ │ lib/ │ │ tokens.py │ ← tiktoken or heuristic │ markdown.py │ ← section parsing │ dedup.py │ ← shingle hashing │ dictionary.py │ ← codebook compression │ rle.py │ ← path/IP/enum encoding │ tokenizer_ │ │ optimizer.py │ ← format optimization │ config.py │ ← JSON config │ exceptions.py │ ← error types └────────────────┘
All commands: python3 scripts/mem_compress.py <workspace> <command> [options] CommandDescriptionTypical SavingsfullComplete pipeline (all steps in order)50%+ combinedbenchmarkDry-run performance report—compressRule-based compression4-8%dictDictionary encoding with auto-codebook4-5%observeSession transcript → observations~97%tiersGenerate L0/L1/L2 summaries88-95% on sub-agent loadsdedupCross-file duplicate detectionvariesestimateToken count report—auditWorkspace health check—optimizeTokenizer-level format fixes1-3%
--json — Machine-readable JSON output --dry-run — Preview changes without writing --since YYYY-MM-DD — Filter sessions by date --auto-merge — Auto-merge duplicates (dedup)
Workspace StateTypical SavingsNotesSession transcripts (observe)~97%Megabytes of JSONL → concise observation MDVerbose/new workspace50-70%First run on unoptimized workspaceRegular maintenance10-20%Weekly runs on active workspaceAlready-optimized3-12%Diminishing returns — workspace is clean
Before compression runs, enable prompt caching for a 90% discount on cached tokens: { "models": { "model-name": { "cacheRetention": "long" } } } Compression reduces token count, caching reduces cost-per-token. Together: 50% compression + 90% cache discount = 95% effective cost reduction.
Optional claw-compactor-config.json in workspace root: { "chars_per_token": 4, "level0_max_tokens": 200, "level1_max_tokens": 500, "dedup_similarity_threshold": 0.6, "dedup_shingle_size": 3 } All fields optional — sensible defaults are used when absent.
FilePurposememory/.codebook.jsonDictionary codebook (must travel with memory files)memory/.observed-sessions.jsonTracks processed transcriptsmemory/observations/Compressed session summariesmemory/MEMORY-L0.mdLevel 0 summary (~200 tokens)
Q: Will compression lose my data? A: Rule engine, dictionary, RLE, and tokenizer optimization are fully lossless. Observation compression and CCP are lossy but preserve all facts and decisions. Q: How does dictionary decompression work? A: decompress_text(text, codebook) expands all $XX codes back. The codebook JSON must be present. Q: Can I run individual steps? A: Yes. Every command is independent: compress, dict, observe, tiers, dedup, optimize. Q: What if tiktoken isn't installed? A: Falls back to a CJK-aware heuristic (chars÷4). Results are ~90% accurate. Q: Does it handle Chinese/Japanese/Unicode? A: Yes. Full CJK support including character-aware token estimation and Chinese punctuation normalization.
FileNotFoundError on workspace: Ensure path points to workspace root (contains memory/ or MEMORY.md) Dictionary decompression fails: Check memory/.codebook.json exists and is valid JSON Zero savings on benchmark: Workspace is already optimized — nothing to do observe finds no transcripts: Check sessions directory for .jsonl files Token count seems wrong: Install tiktoken: pip3 install tiktoken
Inspired by claude-mem by thedotmack Built by Bot777 🤖 for OpenClaw
MIT
Agent frameworks, memory systems, reasoning layers, and model-native orchestration.
Largest current source with strong distribution and engagement signals.