Requirements
- Target platform
- OpenClaw
- Install method
- Manual import
- Extraction
- Extract archive
- Prerequisites
- OpenClaw
- Primary doc
- SKILL.md
Text-to-speech, speech-to-text, voice conversion, and audio processing using EachLabs AI models. Supports ElevenLabs TTS, Whisper transcription with diarization, and RVC voice conversion. Use when the user needs TTS, transcription, or voice conversion.
Text-to-speech, speech-to-text, voice conversion, and audio processing using EachLabs AI models. Supports ElevenLabs TTS, Whisper transcription with diarization, and RVC voice conversion. Use when the user needs TTS, transcription, or voice conversion.
This item's current download entry is known to bounce back to a listing or homepage instead of returning a package file.
Use the source page and any available docs to guide the install because the item currently does not return a direct package file.
I tried to install a skill package from Yavira, but the item currently does not return a direct package file. Inspect the source page and any extracted docs, then tell me what you can confirm and any manual steps still required.
I tried to upgrade a skill package from Yavira, but the item currently does not return a direct package file. Compare the source page and any extracted docs with my current installation, then summarize what changed and what manual follow-up I still need.
Text-to-speech, speech-to-text transcription, voice conversion, and audio utilities via the EachLabs Predictions API.
Header: X-API-Key: <your-api-key> Set the EACHLABS_API_KEY environment variable. Get your key at eachlabs.ai.
ModelSlugBest ForElevenLabs TTSelevenlabs-text-to-speechHigh quality TTSElevenLabs TTS w/ Timestampselevenlabs-text-to-speech-with-timestampTTS with word timingElevenLabs Text to Dialogueelevenlabs-text-to-dialogueMulti-speaker dialogueElevenLabs Sound Effectselevenlabs-sound-effectsSound effect generationElevenLabs Voice Design v2elevenlabs-voice-design-v2Custom voice designKling V1 TTSkling-v1-ttsKling text-to-speechKokoro 82Mkokoro-82mLightweight TTSPlay AI Dialogplay-ai-text-to-speech-dialogDialog TTSStable Audio 2.5stable-audio-2-5-text-to-audioText to audio
ModelSlugBest ForElevenLabs Scribe v2elevenlabs-speech-to-text-scribe-v2Best quality transcriptionElevenLabs STTelevenlabs-speech-to-textStandard transcriptionWizper with Timestampwizper-with-timestampTimestamped transcriptionWizperwizperBasic transcriptionWhisperwhisperOpen-source transcriptionWhisper Diarizationwhisper-diarizationSpeaker identificationIncredibly Fast Whisperincredibly-fast-whisperFastest transcription
ModelSlugBest ForRVC v2rvc-v2Voice conversionTrain RVCtrain-rvcTrain custom voice modelElevenLabs Voice Cloneelevenlabs-voice-cloneVoice cloningElevenLabs Voice Changerelevenlabs-voice-changerVoice transformationElevenLabs Voice Design v3elevenlabs-voice-design-v3Advanced voice designElevenLabs Dubbingelevenlabs-dubbingVideo dubbingChatterbox S2Schatterbox-speech-to-speechSpeech to speechOpen VoiceopenvoiceOpen-source voice cloneXTTS v2xtts-v2Multi-language voice cloneStable Audio 2.5 Inpaintstable-audio-2-5-inpaintAudio inpaintingStable Audio 2.5 A2Astable-audio-2-5-audio-to-audioAudio transformationAudio Trimmeraudio-trimmer-with-fadeAudio trimming with fade
ModelSlugBest ForFFmpeg Merge Audio Videoffmpeg-api-merge-audio-videoMerge audio with videoToolkit Video ConverttoolkitVideo/audio conversion
Check model GET https://api.eachlabs.ai/v1/model?slug=<slug> — validates the model exists and returns the request_schema with exact input parameters. Always do this before creating a prediction to ensure correct inputs. POST https://api.eachlabs.ai/v1/prediction with model slug, version "0.0.1", and input matching the schema Poll GET https://api.eachlabs.ai/v1/prediction/{id} until status is "success" or "failed" Extract the output from the response
curl -X POST https://api.eachlabs.ai/v1/prediction \ -H "Content-Type: application/json" \ -H "X-API-Key: $EACHLABS_API_KEY" \ -d '{ "model": "elevenlabs-text-to-speech", "version": "0.0.1", "input": { "text": "Welcome to our product demo. Today we will walk through the key features.", "voice_id": "EXAVITQu4vr4xnSDxMaL", "model_id": "eleven_v3", "stability": 0.5, "similarity_boost": 0.7 } }'
curl -X POST https://api.eachlabs.ai/v1/prediction \ -H "Content-Type: application/json" \ -H "X-API-Key: $EACHLABS_API_KEY" \ -d '{ "model": "elevenlabs-speech-to-text-scribe-v2", "version": "0.0.1", "input": { "media_url": "https://example.com/recording.mp3", "diarize": true, "timestamps_granularity": "word" } }'
curl -X POST https://api.eachlabs.ai/v1/prediction \ -H "Content-Type: application/json" \ -H "X-API-Key: $EACHLABS_API_KEY" \ -d '{ "model": "wizper-with-timestamp", "version": "0.0.1", "input": { "audio_url": "https://example.com/audio.mp3", "language": "en", "task": "transcribe", "chunk_level": "segment" } }'
curl -X POST https://api.eachlabs.ai/v1/prediction \ -H "Content-Type: application/json" \ -H "X-API-Key: $EACHLABS_API_KEY" \ -d '{ "model": "whisper-diarization", "version": "0.0.1", "input": { "file_url": "https://example.com/meeting.mp3", "num_speakers": 3, "language": "en", "group_segments": true } }'
curl -X POST https://api.eachlabs.ai/v1/prediction \ -H "Content-Type: application/json" \ -H "X-API-Key: $EACHLABS_API_KEY" \ -d '{ "model": "rvc-v2", "version": "0.0.1", "input": { "input_audio": "https://example.com/vocals.wav", "rvc_model": "CUSTOM", "custom_rvc_model_download_url": "https://example.com/my-voice-model.zip", "pitch_change": 0, "output_format": "wav" } }'
curl -X POST https://api.eachlabs.ai/v1/prediction \ -H "Content-Type: application/json" \ -H "X-API-Key: $EACHLABS_API_KEY" \ -d '{ "model": "ffmpeg-api-merge-audio-video", "version": "0.0.1", "input": { "video_url": "https://example.com/video.mp4", "audio_url": "https://example.com/narration.mp3", "start_offset": 0 } }'
The elevenlabs-text-to-speech model supports these voice IDs. Pass the raw ID string: Voice IDNotesEXAVITQu4vr4xnSDxMaLDefault voice9BWtsMINqrJLrRacOk9x—CwhRBWXzGAHq8TQ4Fs17—FGY2WhTYpPnrIDTdsKH5—JBFqnCBsd6RMkjVDRZzb—N2lVS1w4EtoT3dr4eOWO—TX3LPaxmHKxFdv7VOQHJ—XB0fDUnXU5powFXDhCwa—onwK4e9ZLuTAKqWW03F9—pFZP5JQG7iQjIQuC4Bku—
See references/MODELS.md for complete parameter details for each model.
Agent frameworks, memory systems, reasoning layers, and model-native orchestration.
Largest current source with strong distribution and engagement signals.