mirror of
https://github.com/decolua/9router.git
synced 2026-05-08 12:01:28 +00:00
76 lines
2.6 KiB
Markdown
76 lines
2.6 KiB
Markdown
---
|
|
name: 9router-tts
|
|
description: Text-to-speech via 9Router /v1/audio/speech using OpenAI / ElevenLabs / Deepgram / Edge TTS / Google TTS / Hyperbolic / Inworld voices. Use when the user wants to convert text to speech, generate audio, voiceover, narrate, or read text aloud.
|
|
---
|
|
|
|
# 9Router — Text-to-Speech
|
|
|
|
Requires `NINEROUTER_URL` (and `NINEROUTER_KEY` if auth enabled). See https://raw.githubusercontent.com/decolua/9router/main/app/skills/9router/SKILL.md for setup.
|
|
|
|
## Discover voices
|
|
|
|
```bash
|
|
curl $NINEROUTER_URL/v1/models/tts | jq '.data[].id'
|
|
```
|
|
|
|
`model` = voice ID (e.g. `openai/tts-1`, `el/eleven_multilingual_v2`, `edge-tts/en-US-AriaNeural`).
|
|
|
|
## Endpoint
|
|
|
|
`POST $NINEROUTER_URL/v1/audio/speech`
|
|
|
|
| Field | Required | Notes |
|
|
|---|---|---|
|
|
| `model` | yes | voice ID from `/v1/models/tts` |
|
|
| `input` | yes | text to speak |
|
|
|
|
Query `?response_format=mp3` (default, raw bytes) or `?response_format=json` (`{audio: base64, format}`).
|
|
|
|
## Examples
|
|
|
|
Save MP3:
|
|
|
|
```bash
|
|
curl -X POST "$NINEROUTER_URL/v1/audio/speech" \
|
|
-H "Authorization: Bearer $NINEROUTER_KEY" \
|
|
-H "Content-Type: application/json" \
|
|
-d '{"model":"openai/tts-1","input":"Hello world"}' \
|
|
--output speech.mp3
|
|
```
|
|
|
|
JS (save file):
|
|
|
|
```js
|
|
import { writeFile } from "node:fs/promises";
|
|
const r = await fetch(`${process.env.NINEROUTER_URL}/v1/audio/speech`, {
|
|
method: "POST",
|
|
headers: { "Authorization": `Bearer ${process.env.NINEROUTER_KEY}`, "Content-Type": "application/json" },
|
|
body: JSON.stringify({ model: "el/eleven_multilingual_v2", input: "Xin chào" }),
|
|
});
|
|
await writeFile("speech.mp3", Buffer.from(await r.arrayBuffer()));
|
|
```
|
|
|
|
## Response shape
|
|
|
|
Default → raw audio bytes (Content-Type `audio/mp3`).
|
|
|
|
`?response_format=json`:
|
|
```json
|
|
{ "audio": "SUQzBAAAA...", "format": "mp3" }
|
|
```
|
|
|
|
## Provider quirks (model format)
|
|
|
|
| Provider | `model` format | Notes |
|
|
|---|---|---|
|
|
| `openai` | `tts-1/alloy` (model/voice) or just voice | Default model `gpt-4o-mini-tts` |
|
|
| `elevenlabs` | `<model_id>/<voice_id>` or `<voice_id>` | Default model `eleven_flash_v2_5`; list voices in Dashboard |
|
|
| `openrouter` | `openai/gpt-4o-mini-tts/alloy` | Streamed via chat-completions audio modality |
|
|
| `edge-tts` | voice id e.g. `vi-VN-HoaiMyNeural` | **noAuth**; default `vi-VN-HoaiMyNeural` |
|
|
| `google-tts` | language code e.g. `en`, `vi` | **noAuth** |
|
|
| `local-device` | OS voice name (`say -v ?` / SAPI) | **noAuth**; needs `ffmpeg` |
|
|
| `deepgram` | `aura-asteria-en` etc | Token auth |
|
|
| `nvidia`, `inworld`, `cartesia`, `playht` | `model/voice` | Provider-specific auth header |
|
|
| `coqui`, `tortoise` | speaker / voice id | Localhost noAuth |
|
|
| `hyperbolic` | model id | Body = `{text}` only |
|