← Back to library

Sogni Voice 本地 TTS/STT 接入 OpenClaw(Apple Silicon)

解决“想让助手可语音交互但不依赖第三方付费 API”的场景:在 M 系列 Mac 自建语音服务,再让 OpenClaw 走本地 REST 端点。

REDDITDiscovered 2026-02-13Author u/isit2amalready
Prerequisites
  • macOS on Apple Silicon (M1/M2/M3/M4), Homebrew, Node.js are ready.
  • OpenClaw gateway is running and can reach local voice API endpoint.
Steps
  1. Clone `Sogni-AI/sogni-voice` and install dependencies; install system packages `ffmpeg` and `uv` first.
  2. Run interactive setup (`./setup.sh`) and choose enabled engines (Kokoro / Pocket TTS / Qwen3-TTS / Parakeet-MLX).
  3. Start service (`npm run dev` or PM2 mode), then configure OpenClaw voice endpoints to point to local API URL.
  4. Send one voice message and one TTS request in your channel (Telegram/WhatsApp/iMessage) to validate full round-trip.
Commands
brew install ffmpeg uv
npm install
./setup.sh
npm run dev
Verify

OpenClaw can transcribe uploaded audio and return synthesized voice from local endpoint with no third-party TTS API calls.

Caveats
  • Project currently targets Apple Silicon only; Intel/other platforms need alternative stacks(需验证)。
  • Do not expose unauthenticated voice endpoint to public network.
Source attribution

This tip is aggregated from community/public sources and preserved with attribution.

Open original source ↗
Visit original post