Requirements
- Target platform
- OpenClaw
- Install method
- Manual import
- Extraction
- Extract archive
- Prerequisites
- OpenClaw
- Primary doc
- SKILL.md
Identifies and tracks speakers in multi-person conversations, mapping speaker labels to names and managing voice command authorization levels.
Identifies and tracks speakers in multi-person conversations, mapping speaker labels to names and managing voice command authorization levels.
Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.
I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Tell me what you changed and call out any manual steps you could not complete.
I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Summarize what changed and any follow-up checks I should run.
Speaker identification and management for multi-person conversations.
Tracks who said what in conversations. Maps anonymous speaker labels (SPEAKER_0, SPEAKER_1) to real names, maintains speaker profiles, and gates voice command authorization.
User asks "who said that?" or wants speaker-attributed transcripts User wants to configure which people can trigger voice commands Agent needs to know who is speaking in a multi-person conversation
percept-listen skill installed and running Omi pendant (provides is_user flag for primary speaker)
Omi sends transcript segments with speaker labels (SPEAKER_0, SPEAKER_1, etc.) Percept resolves labels to names using the speakers registry is_user flag from Omi identifies the pendant wearer as the primary speaker Speaker profiles track first/last seen timestamps and authorization status
Located at percept/data/speakers.json: { "SPEAKER_00": { "name": "David", "is_owner": true, "approved": true }, "SPEAKER_01": { "name": "Rob", "is_owner": false, "approved": true } } Manage via Percept dashboard (port 8960) β Settings β Speakers.
Owner (is_owner: true): Full command access, always authorized Approved (approved: true): Can trigger wake word commands Unknown: Logged only, commands not executed
Planned: pyannote speaker diarization with 192-dim voice embeddings for automatic speaker recognition via cosine similarity. Currently speaker mapping is manual.
GitHub: https://github.com/GetPercept/percept
Agent frameworks, memory systems, reasoning layers, and model-native orchestration.
Largest current source with strong distribution and engagement signals.