Requirements
- Target platform
- OpenClaw
- Install method
- Manual import
- Extraction
- Extract archive
- Prerequisites
- OpenClaw
- Primary doc
- SKILL.md
Text-to-speech, speech-to-text, voice conversion, and audio processing using EachLabs AI models. Supports ElevenLabs TTS, Whisper transcription with diarization, and RVC voice conversion. Use when the user needs TTS, transcription, or voice conversion.
Text-to-speech, speech-to-text, voice conversion, and audio processing using EachLabs AI models. Supports ElevenLabs TTS, Whisper transcription with diarization, and RVC voice conversion. Use when the user needs TTS, transcription, or voice conversion.
Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.
I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Tell me what you changed and call out any manual steps you could not complete.
I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Summarize what changed and any follow-up checks I should run.
Text-to-speech, speech-to-text transcription, voice conversion, and audio utilities via the EachLabs Predictions API.
Header: X-API-Key: <your-api-key> Set the EACHLABS_API_KEY environment variable. Get your key at eachlabs.ai.
ModelSlugBest ForElevenLabs TTSelevenlabs-text-to-speechHigh quality TTSElevenLabs TTS w/ Timestampselevenlabs-text-to-speech-with-timestampTTS with word timingElevenLabs Text to Dialogueelevenlabs-text-to-dialogueMulti-speaker dialogueElevenLabs Sound Effectselevenlabs-sound-effectsSound effect generationElevenLabs Voice Design v2elevenlabs-voice-design-v2Custom voice designKling V1 TTSkling-v1-ttsKling text-to-speechKokoro 82Mkokoro-82mLightweight TTSPlay AI Dialogplay-ai-text-to-speech-dialogDialog TTSStable Audio 2.5stable-audio-2-5-text-to-audioText to audio
ModelSlugBest ForElevenLabs Scribe v2elevenlabs-speech-to-text-scribe-v2Best quality transcriptionElevenLabs STTelevenlabs-speech-to-textStandard transcriptionWizper with Timestampwizper-with-timestampTimestamped transcriptionWizperwizperBasic transcriptionWhisperwhisperOpen-source transcriptionWhisper Diarizationwhisper-diarizationSpeaker identificationIncredibly Fast Whisperincredibly-fast-whisperFastest transcription
ModelSlugBest ForRVC v2rvc-v2Voice conversionTrain RVCtrain-rvcTrain custom voice modelElevenLabs Voice Cloneelevenlabs-voice-cloneVoice cloningElevenLabs Voice Changerelevenlabs-voice-changerVoice transformationElevenLabs Voice Design v3elevenlabs-voice-design-v3Advanced voice designElevenLabs Dubbingelevenlabs-dubbingVideo dubbingChatterbox S2Schatterbox-speech-to-speechSpeech to speechOpen VoiceopenvoiceOpen-source voice cloneXTTS v2xtts-v2Multi-language voice cloneStable Audio 2.5 Inpaintstable-audio-2-5-inpaintAudio inpaintingStable Audio 2.5 A2Astable-audio-2-5-audio-to-audioAudio transformationAudio Trimmeraudio-trimmer-with-fadeAudio trimming with fade
ModelSlugBest ForFFmpeg Merge Audio Videoffmpeg-api-merge-audio-videoMerge audio with videoToolkit Video ConverttoolkitVideo/audio conversion
Check model GET https://api.eachlabs.ai/v1/model?slug=<slug> — validates the model exists and returns the request_schema with exact input parameters. Always do this before creating a prediction to ensure correct inputs. POST https://api.eachlabs.ai/v1/prediction with model slug, version "0.0.1", and input matching the schema Poll GET https://api.eachlabs.ai/v1/prediction/{id} until status is "success" or "failed" Extract the output from the response
curl -X POST https://api.eachlabs.ai/v1/prediction \ -H "Content-Type: application/json" \ -H "X-API-Key: $EACHLABS_API_KEY" \ -d '{ "model": "elevenlabs-text-to-speech", "version": "0.0.1", "input": { "text": "Welcome to our product demo. Today we will walk through the key features.", "voice_id": "EXAVITQu4vr4xnSDxMaL", "model_id": "eleven_v3", "stability": 0.5, "similarity_boost": 0.7 } }'
curl -X POST https://api.eachlabs.ai/v1/prediction \ -H "Content-Type: application/json" \ -H "X-API-Key: $EACHLABS_API_KEY" \ -d '{ "model": "elevenlabs-speech-to-text-scribe-v2", "version": "0.0.1", "input": { "media_url": "https://example.com/recording.mp3", "diarize": true, "timestamps_granularity": "word" } }'
curl -X POST https://api.eachlabs.ai/v1/prediction \ -H "Content-Type: application/json" \ -H "X-API-Key: $EACHLABS_API_KEY" \ -d '{ "model": "wizper-with-timestamp", "version": "0.0.1", "input": { "audio_url": "https://example.com/audio.mp3", "language": "en", "task": "transcribe", "chunk_level": "segment" } }'
curl -X POST https://api.eachlabs.ai/v1/prediction \ -H "Content-Type: application/json" \ -H "X-API-Key: $EACHLABS_API_KEY" \ -d '{ "model": "whisper-diarization", "version": "0.0.1", "input": { "file_url": "https://example.com/meeting.mp3", "num_speakers": 3, "language": "en", "group_segments": true } }'
curl -X POST https://api.eachlabs.ai/v1/prediction \ -H "Content-Type: application/json" \ -H "X-API-Key: $EACHLABS_API_KEY" \ -d '{ "model": "rvc-v2", "version": "0.0.1", "input": { "input_audio": "https://example.com/vocals.wav", "rvc_model": "CUSTOM", "custom_rvc_model_download_url": "https://example.com/my-voice-model.zip", "pitch_change": 0, "output_format": "wav" } }'
curl -X POST https://api.eachlabs.ai/v1/prediction \ -H "Content-Type: application/json" \ -H "X-API-Key: $EACHLABS_API_KEY" \ -d '{ "model": "ffmpeg-api-merge-audio-video", "version": "0.0.1", "input": { "video_url": "https://example.com/video.mp4", "audio_url": "https://example.com/narration.mp3", "start_offset": 0 } }'
The elevenlabs-text-to-speech model supports these voice IDs. Pass the raw ID string: Voice IDNotesEXAVITQu4vr4xnSDxMaLDefault voice9BWtsMINqrJLrRacOk9x—CwhRBWXzGAHq8TQ4Fs17—FGY2WhTYpPnrIDTdsKH5—JBFqnCBsd6RMkjVDRZzb—N2lVS1w4EtoT3dr4eOWO—TX3LPaxmHKxFdv7VOQHJ—XB0fDUnXU5powFXDhCwa—onwK4e9ZLuTAKqWW03F9—pFZP5JQG7iQjIQuC4Bku—
See references/MODELS.md for complete parameter details for each model.
Agent frameworks, memory systems, reasoning layers, and model-native orchestration.
Largest current source with strong distribution and engagement signals.