Requirements
- Target platform
- OpenClaw
- Install method
- Manual import
- Extraction
- Extract archive
- Prerequisites
- OpenClaw
- Primary doc
- SKILL.md
Orchestrate multiple Claude Code workers via iTerm2 using the claude-team MCP server. Spawn workers with git worktrees, assign beads issues, monitor progress, and coordinate parallel development work.
Orchestrate multiple Claude Code workers via iTerm2 using the claude-team MCP server. Spawn workers with git worktrees, assign beads issues, monitor progress, and coordinate parallel development work.
This item's current download entry is known to bounce back to a listing or homepage instead of returning a package file.
Use the source page and any available docs to guide the install because the item currently does not return a direct package file.
I tried to install a skill package from Yavira, but the item currently does not return a direct package file. Inspect the source page and any extracted docs, then tell me what you can confirm and any manual steps still required.
I tried to upgrade a skill package from Yavira, but the item currently does not return a direct package file. Compare the source page and any extracted docs with my current installation, then summarize what changed and what manual follow-up I still need.
Claude-team is an MCP server that lets you spawn and manage teams of Claude Code sessions via iTerm2. Each worker gets their own terminal pane, optional git worktree, and can be assigned beads issues.
Parallelism: Fan out work to multiple agents working simultaneously Context isolation: Each worker has fresh context, keeps coordinator context clean Visibility: Real Claude Code sessions you can watch, interrupt, or take over Git worktrees: Each worker can have an isolated branch for their work
NEVER make code changes directly. Always spawn workers for code changes. This keeps your context clean and provides proper git workflow with worktrees.
macOS with iTerm2 (Python API enabled: Preferences โ General โ Magic โ Enable Python API) claude-team MCP server configured in ~/.claude.json
All tools are called through mcporter call claude-team.<tool>: mcporter call claude-team.list_workers mcporter call claude-team.spawn_workers workers='[{"project_path":"/path/to/repo","bead":"cp-123"}]'
Create new Claude Code worker sessions. mcporter call claude-team.spawn_workers \ workers='[{ "project_path": "/path/to/repo", "bead": "cp-123", "annotation": "Fix auth bug", "use_worktree": true, "skip_permissions": true }]' \ layout="auto" Worker config fields: project_path: Required. Path to repo or "auto" (uses CLAUDE_TEAM_PROJECT_DIR) bead: Optional beads issue ID โ worker will follow beads workflow annotation: Task description (shown on badge, used in branch name) prompt: Additional instructions (if no bead, this is their assignment) use_worktree: Create isolated git worktree (default: true) skip_permissions: Start with --dangerously-skip-permissions (default: false) name: Optional worker name override (auto-picks from themed sets otherwise) Layout options: "auto": Reuse existing claude-team windows, split into available space "new": Always create fresh window (1-4 workers in grid layout)
See all managed workers: mcporter call claude-team.list_workers mcporter call claude-team.list_workers status_filter="ready" Status values: spawning, ready, busy, closed
Send messages to one or more workers: mcporter call claude-team.message_workers \ session_ids='["Groucho"]' \ message="Please also add unit tests" \ wait_mode="none" wait_mode options: "none": Fire and forget (default) "any": Return when any worker is idle "all": Return when all workers are idle
Check or wait for workers to finish: # Quick poll mcporter call claude-team.check_idle_workers session_ids='["Groucho","Harpo"]' # Blocking wait mcporter call claude-team.wait_idle_workers \ session_ids='["Groucho","Harpo"]' \ mode="all" \ timeout=600
Get conversation history: mcporter call claude-team.read_worker_logs \ session_id="Groucho" \ pages=2
Get detailed status including conversation stats: mcporter call claude-team.examine_worker session_id="Groucho"
Terminate workers when done: mcporter call claude-team.close_workers session_ids='["Groucho","Harpo"]' โ ๏ธ Worktree cleanup: Workers with worktrees commit to ephemeral branches. After closing: Review commits on the worker's branch Merge or cherry-pick to a persistent branch Delete the branch: git branch -D <branch-name>
Quick reference for beads commands: mcporter call claude-team.bd_help
Workers can be referenced by any of: Internal ID: Short hex string (e.g., 3962c5c4) Terminal ID: iterm:UUID format Worker name: Human-friendly name (e.g., Groucho, Aragorn)
# 1. Spawn worker with a bead assignment mcporter call claude-team.spawn_workers \ workers='[{ "project_path": "/Users/phaedrus/Projects/myrepo", "bead": "proj-abc", "annotation": "Implement config schemas", "use_worktree": true, "skip_permissions": true }]' # 2. Worker automatically: # - Creates worktree with branch named after bead # - Runs `bd show proj-abc` to understand the task # - Marks issue in_progress # - Implements the work # - Closes the issue # - Commits with issue reference # 3. Monitor progress mcporter call claude-team.check_idle_workers session_ids='["Groucho"]' mcporter call claude-team.read_worker_logs session_id="Groucho" # 4. When done, close and merge mcporter call claude-team.close_workers session_ids='["Groucho"]' # Then: git merge or cherry-pick from worker's branch
# Spawn multiple workers for parallel tasks mcporter call claude-team.spawn_workers \ workers='[ {"project_path": "auto", "bead": "cp-123", "annotation": "Auth module"}, {"project_path": "auto", "bead": "cp-124", "annotation": "API routes"}, {"project_path": "auto", "bead": "cp-125", "annotation": "Unit tests"} ]' \ layout="new" # Wait for all to complete mcporter call claude-team.wait_idle_workers \ session_ids='["Groucho","Harpo","Chico"]' \ mode="all" # Review and close mcporter call claude-team.close_workers \ session_ids='["Groucho","Harpo","Chico"]'
Use beads: Assign bead IDs so workers follow proper issue workflow Use worktrees: Keeps work isolated, enables parallel commits Skip permissions: Workers need skip_permissions: true to write files Monitor, don't micromanage: Let workers complete, then review Merge carefully: Review worker branches before merging to main Close workers: Always close when done to clean up worktrees
For persistent server operation, claude-team can run as an HTTP server. This keeps the MCP server running continuously with persistent state, avoiding cold starts.
Run the claude-team HTTP server directly: # From the claude-team directory uv run python -m claude_team_mcp --http --port 8766 # Or specify the directory explicitly uv run --directory /path/to/claude-team python -m claude_team_mcp --http --port 8766 For automatic startup on login, use launchd (see the "launchd Auto-Start" section below).
Once the HTTP server is running, configure mcporter to connect to it. Create ~/.mcporter/mcporter.json: { "mcpServers": { "claude-team": { "transport": "streamable-http", "url": "http://127.0.0.1:8766/mcp", "lifecycle": "keep-alive" } } }
Persistent state: Worker registry survives across CLI invocations Faster responses: No Python environment startup on each call External access: Can be accessed by cron jobs, scripts, or other tools Session recovery: Server tracks sessions even if coordinator disconnects
Update your .mcp.json to use HTTP transport: { "mcpServers": { "claude-team": { "transport": "streamable-http", "url": "http://127.0.0.1:8766/mcp" } } }
To automatically start the claude-team server on login, use the bundled setup script.
Run the setup script from the skill's assets directory: # From the skill directory ./assets/setup.sh # Or specify a custom claude-team location CLAUDE_TEAM_DIR=/path/to/claude-team ./assets/setup.sh
The setup script: Detects your uv installation path Creates the log directory at ~/.claude-team/logs/ Generates a launchd plist from assets/com.claude-team.plist.template Installs it to ~/Library/LaunchAgents/com.claude-team.plist Loads the service to start immediately The plist template uses uv run to start the HTTP server on port 8766, configured for iTerm2 Python API access (Aqua session type).
# Stop the service launchctl unload ~/Library/LaunchAgents/com.claude-team.plist # Restart (re-run setup) ./assets/setup.sh # Check if running launchctl list | grep claude-team # View logs tail -f ~/.claude-team/logs/stdout.log tail -f ~/.claude-team/logs/stderr.log
# Check for load errors launchctl print gui/$UID/com.claude-team # Force restart launchctl kickstart -k gui/$UID/com.claude-team # Remove and reload (if plist changed) launchctl bootout gui/$UID/com.claude-team launchctl bootstrap gui/$UID ~/Library/LaunchAgents/com.claude-team.plist
For background monitoring and notifications, claude-team supports cron-based worker tracking.
Claude-team writes worker state to ~/.claude-team/memory/worker-tracking.json: { "workers": { "Groucho": { "session_id": "3962c5c4", "bead": "cp-123", "annotation": "Fix auth bug", "status": "busy", "project_path": "/Users/phaedrus/Projects/myrepo", "started_at": "2025-01-05T10:30:00Z", "last_activity": "2025-01-05T11:45:00Z" }, "Harpo": { "session_id": "a1b2c3d4", "bead": "cp-124", "annotation": "Add API routes", "status": "idle", "project_path": "/Users/phaedrus/Projects/myrepo", "started_at": "2025-01-05T10:30:00Z", "last_activity": "2025-01-05T11:50:00Z", "completed_at": "2025-01-05T11:50:00Z" } }, "last_updated": "2025-01-05T11:50:00Z" }
Create a monitoring script at ~/.claude-team/scripts/check-workers.sh: #!/bin/bash # Check for completed workers and send notifications TRACKING_FILE="$HOME/.claude-team/memory/worker-tracking.json" NOTIFIED_FILE="$HOME/.claude-team/memory/notified-workers.json" TELEGRAM_BOT_TOKEN="${TELEGRAM_BOT_TOKEN}" TELEGRAM_CHAT_ID="${TELEGRAM_CHAT_ID}" # Exit if tracking file doesn't exist [ -f "$TRACKING_FILE" ] || exit 0 # Initialize notified file if needed [ -f "$NOTIFIED_FILE" ] || echo '{"notified":[]}' > "$NOTIFIED_FILE" # Find idle workers that haven't been notified IDLE_WORKERS=$(jq -r ' .workers | to_entries[] | select(.value.status == "idle") | .key ' "$TRACKING_FILE") for worker in $IDLE_WORKERS; do # Check if already notified ALREADY_NOTIFIED=$(jq -r --arg w "$worker" '.notified | index($w) != null' "$NOTIFIED_FILE") if [ "$ALREADY_NOTIFIED" = "false" ]; then # Get worker details BEAD=$(jq -r --arg w "$worker" '.workers[$w].bead // "no-bead"' "$TRACKING_FILE") ANNOTATION=$(jq -r --arg w "$worker" '.workers[$w].annotation // "no annotation"' "$TRACKING_FILE") # Send Telegram notification MESSAGE="๐ค Worker *${worker}* completed ๐ Bead: \`${BEAD}\` ๐ ${ANNOTATION}" curl -s -X POST "https://api.telegram.org/bot${TELEGRAM_BOT_TOKEN}/sendMessage" \ -d chat_id="$TELEGRAM_CHAT_ID" \ -d text="$MESSAGE" \ -d parse_mode="Markdown" > /dev/null # Mark as notified jq --arg w "$worker" '.notified += [$w]' "$NOTIFIED_FILE" > "${NOTIFIED_FILE}.tmp" mv "${NOTIFIED_FILE}.tmp" "$NOTIFIED_FILE" fi done Make it executable: chmod +x ~/.claude-team/scripts/check-workers.sh
Add to crontab (crontab -e): # Check claude-team workers every 2 minutes */2 * * * * TELEGRAM_BOT_TOKEN="your-bot-token" TELEGRAM_CHAT_ID="your-chat-id" ~/.claude-team/scripts/check-workers.sh
Set Telegram credentials in your shell profile (~/.zshrc): export TELEGRAM_BOT_TOKEN="123456789:ABCdefGHIjklMNOpqrsTUVwxyz" export TELEGRAM_CHAT_ID="-1001234567890"
If you have clawdbot configured, you can send notifications through it instead: # In check-workers.sh, replace the curl command with: clawdbot send --to "$TELEGRAM_CHAT_ID" --message "$MESSAGE" --provider telegram
When starting a fresh batch of workers, clear the notified list: echo '{"notified":[]}' > ~/.claude-team/memory/notified-workers.json
Code helpers, APIs, CLIs, browser automation, testing, and developer operations.
Largest current source with strong distribution and engagement signals.