Human setup (one-time)
- Create or edit
~/.openclaw/.envon the machine running OpenClaw. - Add your credentials (do not commit these):
ELEVENLABS_API_KEY=your_elevenlabs_key
# Optional (recommended): default voice
ELEVENLABS_VOICE_ID=your_voice_id
- Restart the OpenClaw gateway (so the service environment picks up changes).
Inputs to collect (if missing)
- Text to speak.
- Optional:
voiceId(ElevenLabs voice id). If not provided, use the default. - Optional: output format (
mp3default).
Requirements (credentials)
This skill does not embed secrets. Credentials must be provided at runtime:
ELEVENLABS_API_KEY(required)- Recommended: put it in
~/.openclaw/.envon the machine running the gateway.
- Recommended: put it in
ELEVENLABS_VOICE_ID(optional but recommended)- If omitted, you must pass
--voice-idwhen calling the script.
- If omitted, you must pass
Example user prompts (to trigger this skill)
- "Reply with a voice note saying: …"
- "Generate audio (ElevenLabs) for: …"
- "Envíame una nota de voz que diga: …"
Workflow
- Choose output path under the workspace, e.g.
./media/elevenlabs-tts/<timestamp>.mp3. - Run:
python3 {baseDir}/scripts/tts.py --text "..." --out "/abs/path/to/file.mp3" [--voice-id "..."]
- Confirm with a short message including the output file path.
Defaults
- No voice is hardcoded. Set
ELEVENLABS_VOICE_ID(recommended) or pass--voice-id. - Model:
eleven_multilingual_v2.