Requirements
- Target platform
- OpenClaw
- Install method
- Manual import
- Extraction
- Extract archive
- Prerequisites
- OpenClaw
- Primary doc
- SKILL.md
AI agent governance, trust scoring, and policy enforcement powered by AgentMesh. Activate when: (1) user wants to enforce token limits, tool restrictions, or...
AI agent governance, trust scoring, and policy enforcement powered by AgentMesh. Activate when: (1) user wants to enforce token limits, tool restrictions, or...
Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.
I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Tell me what you changed and call out any manual steps you could not complete.
I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Summarize what changed and any follow-up checks I should run.
Zero-trust governance layer for OpenClaw agents. Enforce policies, verify identities, score trust, and maintain tamper-evident audit logs โ all from your agent's command line.
Install the AgentMesh governance CLI: pip install agentmesh-governance If agentmesh-governance is not yet on PyPI, install directly from source: pip install "agentmesh @ git+https://github.com/imran-siddique/agent-mesh.git"
All scripts are in scripts/. They wrap the governance engine and output JSON results.
Evaluate an action against a governance policy before execution: scripts/check-policy.sh --action "web_search" --tokens 1500 --policy policy.yaml Returns JSON with allowed: true/false, any violations, and recommendations. Use this before executing any tool call to enforce limits.
Check an agent's current trust score (0.0 โ 1.0): scripts/trust-score.sh --agent "research-agent" Returns the composite trust score with breakdown across 5 dimensions: policy compliance, resource efficiency, output quality, security posture, collaboration health.
Verify an agent's Ed25519 cryptographic identity before trusting its output: scripts/verify-identity.sh --did "did:agentmesh:abc123" --message "hello" --signature "base64sig" Returns verified: true/false. Use when receiving data from another agent.
Update trust scores after collaborating with another agent: scripts/record-interaction.sh --agent "writer-agent" --outcome success scripts/record-interaction.sh --agent "writer-agent" --outcome failure --severity 0.1 Success adds +0.01 to trust score. Failure subtracts the severity value. Agents dropping below the minimum threshold (default 0.5) are auto-blocked.
View tamper-evident audit trail with Merkle chain verification: scripts/audit-log.sh --last 20 scripts/audit-log.sh --agent "research-agent" --verify The --verify flag checks Merkle chain integrity โ any tampering is detected.
Create a new Ed25519 cryptographic identity (DID) for your agent: scripts/generate-identity.sh --name "my-agent" --capabilities "search,summarize,write" Returns your agent's DID, public key, and capability manifest.
Create a policy.yaml to define governance rules: name: production-policy max_tokens: 4096 max_tool_calls: 10 allowed_tools: - web_search - file_read - summarize blocked_tools: - shell_exec - file_delete blocked_patterns: - "rm -rf" - "DROP TABLE" - "BEGIN CERTIFICATE" confidence_threshold: 0.7 require_human_approval: false
Before tool execution: Run check-policy.sh to enforce limits Before trusting another agent's output: Run verify-identity.sh After collaboration: Run record-interaction.sh to update trust Before delegation: Check trust-score.sh โ don't delegate to agents below 0.5 For compliance: Run audit-log.sh --verify to prove execution integrity On setup: Run generate-identity.sh to create your agent's DID
PolicyDescriptionToken limitsCap per-action and per-session token usageTool allowlistsOnly explicitly permitted tools can executeTool blocklistsDangerous tools are blocked regardlessContent patternsBlock regex patterns (secrets, destructive commands, PII)Trust thresholdsMinimum trust score required for delegationHuman approvalGate critical actions behind human confirmation
This skill bridges the OpenClaw agent runtime with the AgentMesh governance engine: OpenClaw Agent โ SKILL.md scripts โ AgentMesh Engine โโโ GovernancePolicy (enforcement) โโโ TrustEngine (5-dimension scoring) โโโ AgentIdentity (Ed25519 DIDs) โโโ MerkleAuditChain (tamper-evident logs) Part of the Agent Ecosystem: AgentMesh ยท Agent OS ยท Agent SRE
Long-tail utilities that do not fit the current primary taxonomy cleanly.
Largest current source with strong distribution and engagement signals.