Sogni Voice 本地 TTS/STT 接入 OpenClaw(Apple Silicon)
解决“想让助手可语音交互但不依赖第三方付费 API”的场景:在 M 系列 Mac 自建语音服务,再让 OpenClaw 走本地 REST 端点。
REDDITDiscovered 2026-02-13Author u/isit2amalready
Prerequisites
- macOS on Apple Silicon (M1/M2/M3/M4), Homebrew, Node.js are ready.
- OpenClaw gateway is running and can reach local voice API endpoint.
Steps
- Clone `Sogni-AI/sogni-voice` and install dependencies; install system packages `ffmpeg` and `uv` first.
- Run interactive setup (`./setup.sh`) and choose enabled engines (Kokoro / Pocket TTS / Qwen3-TTS / Parakeet-MLX).
- Start service (`npm run dev` or PM2 mode), then configure OpenClaw voice endpoints to point to local API URL.
- Send one voice message and one TTS request in your channel (Telegram/WhatsApp/iMessage) to validate full round-trip.
Commands
brew install ffmpeg uvnpm install./setup.shnpm run devVerify
OpenClaw can transcribe uploaded audio and return synthesized voice from local endpoint with no third-party TTS API calls.
Caveats
- Project currently targets Apple Silicon only; Intel/other platforms need alternative stacks(需验证)。
- Do not expose unauthenticated voice endpoint to public network.
Source attribution
This tip is aggregated from community/public sources and preserved with attribution.
Open original source ↗