Requirements
- Target platform
- OpenClaw
- Install method
- Manual import
- Extraction
- Extract archive
- Prerequisites
- OpenClaw
- Primary doc
- SKILL.md
Built by Axobotl (@Inner_Axiom). Production-grade Honcho memory system battle-tested on a 6+ agent fleet with 1000+ messages. Replaces embedding-based retrie...
Built by Axobotl (@Inner_Axiom). Production-grade Honcho memory system battle-tested on a 6+ agent fleet with 1000+ messages. Replaces embedding-based retrie...
Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.
I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Tell me what you changed and call out any manual steps you could not complete.
I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Summarize what changed and any follow-up checks I should run.
Production memory system for AI agents, built on Honcho by Plastic Labs. Developed and refined running a fleet of 6+ agents in production with 1000+ messages fed through the reasoning engine. This isn't a wrapper or setup guide. It's a complete memory pipeline: feed conversations in, reason over them automatically, generate token-budgeted context files, and query on-demand when agents need to recall something they've never seen in their current session.
Agent memory is broken in three ways: Compaction amnesia. Long sessions get compacted. Context vanishes. Your agent forgets decisions made 2 hours ago. Isolated session blindness. Cron jobs and background tasks spin up fresh sessions with zero conversation history. They operate without context. Single-agent silos. In multi-agent setups, agents can't access what other agents learned. Knowledge stays trapped in individual sessions. Honcho Memory solves all three. Every agent feeds into a shared reasoning engine. Every session (main, cron, isolated) loads reasoned context at startup. Nothing gets lost.
Most agent memory is embedding search: store text chunks, retrieve similar chunks later. That's a library with a search bar. Honcho is a reasoning engine. It processes conversations through Neuromancer and builds deductive and inductive observations about every person and agent it tracks. It doesn't retrieve what was said. It reasons about what it means. Feed it 3 weeks of multi-agent conversations and ask "What should the social media agent focus on this week?" It won't dump old messages. It will synthesize: "Based on recent engagement patterns, sports/entertainment replies are outperforming crypto content 3:1. The user shifted priority toward brand-building over direct promotion last Tuesday."
WRITE PATH Conversations โโโ feed.py โโโ Honcho API โโโ Neuromancer โโโ Observations (all agents) โ (async reasoning) โ โ โ โ Representations โ (per peer) โ โ โ READ PATH โ โ โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ โ โ โ generate-context.py (cron) โโโโโโโโโโโโโโ โ โ Token-budgeted files that โ โ โ survive compaction โ โ โโโโโโโโโโโโฌโโโโโโโโโโโโโโโโโโโ โ โ โ โ โ HONCHO-CONTEXT.md (shared) โ agents/*/HONCHO-CONTEXT.md (per-agent) โ โ โ โ โ Every session loads reasoned context โ (main, cron, isolated, sub-agent) โ โ โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ โโโโโโ query.py (on-demand) โโโโ Agent mid-conversation โ "What did we decide about โ needs to recall something โ pricing last week?" โ โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
Context files are workspace files. They survive compaction, session resets, and gateway restarts. Every session starts with reasoned context automatically, not just the last few messages.
All agents feed into the same Honcho workspace. Agent A learns something on Monday. Agent B can access that understanding on Tuesday through shared context generation. No manual syncing, no copy-pasting between workspaces.
Generated context files are budget-controlled (~1000-1500 tokens shared, ~500-800 per agent). No prompt bloat. Honcho distills weeks of conversations into dense, relevant context that fits cleanly in the system prompt.
On-demand queries support five reasoning levels: minimal, low, medium, high, max. Quick lookups stay cheap. Deep analysis gets thorough treatment.
Designed to complement existing file-based memory (MEMORY.md, memory/*.md), not replace it. Honcho handles reasoned recall. Local files handle raw logs and structured data. If Honcho is unavailable, file-based memory still works.
Python 3.10+ Honcho account + API key (app.honcho.dev) pip install honcho-ai
pip install honcho-ai
Create ~/.config/honcho/credentials.json: { "api_key": "your-honcho-api-key", "workspace_id": "your-workspace-name" }
python3 scripts/setup.py --workspace "my-workspace" --peers "user,agent1,agent2" Creates the workspace and registers every entity (users and agents) that Honcho will track and build representations for.
python3 scripts/feed.py Auto-discovers active agent session transcripts, attributes messages to correct peers, tracks sync position to avoid duplicates, feeds in batches. Run after sessions or on a cron schedule.
When an agent needs context it doesn't have: python3 scripts/query.py "What was the decision on pricing?" python3 scripts/query.py "What are the user's priorities right now?" --peer user-name python3 scripts/query.py "Full analysis of this week's performance" --level high
python3 scripts/generate-context.py # all agents python3 scripts/generate-context.py --agent agent-name # specific agent python3 scripts/generate-context.py --query "topic" # on-demand deep query Produces: HONCHO-CONTEXT.md in workspace root (shared across all agents) HONCHO-CONTEXT.md in each agent's workspace (agent-specific)
python3 scripts/feed.py --status
Run hourly during active hours to keep memory fresh: 0 9-23 * * * cd /workspace && python3 scripts/feed.py && python3 scripts/generate-context.py OpenClaw cron tool config: { "name": "honcho-sync", "schedule": {"kind": "cron", "expr": "0 9-23 * * *", "tz": "America/Los_Angeles"}, "sessionTarget": "isolated", "payload": { "kind": "agentTurn", "message": "Run: cd /workspace && source .venv/bin/activate && python3 scripts/feed.py && python3 scripts/generate-context.py. Report summary.", "timeoutSeconds": 180 }, "delivery": {"mode": "none"} }
For fleets with multiple agents, each agent gets: Its own peer in Honcho (tracks what that agent said and learned) Shared context from the root HONCHO-CONTEXT.md (cross-agent awareness) Agent-specific context from its workspace HONCHO-CONTEXT.md (role-specific memory) The feed script handles attribution automatically. User messages go to the human peer. Each agent's responses go to that agent's peer. Honcho reasons across all of them. Example: 6-agent fleet Main agent, social media agent, content agent, community agent, news scanner, expansion agent All feed into one Honcho workspace Each gets tailored context files reflecting their role and recent activity The main agent's context includes a cross-agent brief summarizing what every other agent did
Cron job needs context: Generated context files give every isolated session full awareness of recent activity, decisions, and priorities without carrying conversation history. Agent forgot a decision: On-demand query pulls reasoned context about any topic from weeks of conversations. New agent joins the fleet: Reads shared HONCHO-CONTEXT.md and immediately understands the user, active projects, and current priorities. Post-compaction recovery: Context files survive compaction. Agent reloads and continues without losing the thread. Cross-agent coordination: Social media agent needs to know what the content agent published. Shared context provides this automatically.
Honcho's data model: Workspaces โ Peers โ Sessions โ Messages. Workspaces isolate environments. Peers are any entity that persists and changes over time (users, agents). Sessions are conversation threads between peers. Messages are units of data that trigger reasoning. When messages are written, Honcho doesn't just store them. Custom reasoning models (Neuromancer) perform formal logical reasoning in the background to generate conclusions about each peer. These conclusions compound into representations that evolve with every new interaction. When you query, you get reasoned context derived from representations, not raw message retrieval. Observations: Deductive and inductive conclusions drawn from conversations Representations: Synthesized, evolving understanding of each peer Context retrieval: Token-budgeted, combining representations, recent activity, and reasoned observations For full details on Honcho's architecture: docs.honcho.dev Cost: ~$2/million tokens ingested. Storage and retrieval free. Most multi-agent setups run $4-6/month.
references/api-reference.md โ Full Honcho SDK API documentation references/architecture.md โ Integration patterns and data flow diagrams Honcho docs Plastic Labs
Messaging, meetings, inboxes, CRM, and teammate communication surfaces.
Largest current source with strong distribution and engagement signals.