430 lines
11 KiB
Markdown
430 lines
11 KiB
Markdown
|
|
# WebSocket Streaming API Documentation
|
|||
|
|
|
|||
|
|
## Overview
|
|||
|
|
|
|||
|
|
The WebSocket endpoint (`/ws/stream`) enables real-time, token-by-token TTS streaming, perfect for Discord voice chat with LLM streaming responses.
|
|||
|
|
|
|||
|
|
## Endpoint
|
|||
|
|
|
|||
|
|
```
|
|||
|
|
ws://localhost:8765/ws/stream
|
|||
|
|
```
|
|||
|
|
|
|||
|
|
## Protocol
|
|||
|
|
|
|||
|
|
### Client → Server Messages
|
|||
|
|
|
|||
|
|
**Send Token:**
|
|||
|
|
```json
|
|||
|
|
{
|
|||
|
|
"token": "Hello",
|
|||
|
|
"pitch_shift": 0
|
|||
|
|
}
|
|||
|
|
```
|
|||
|
|
|
|||
|
|
**Flush Buffer:**
|
|||
|
|
```json
|
|||
|
|
{
|
|||
|
|
"flush": true
|
|||
|
|
}
|
|||
|
|
```
|
|||
|
|
|
|||
|
|
### Server → Client Messages
|
|||
|
|
|
|||
|
|
**Audio Data (Binary):**
|
|||
|
|
- Format: PCM float32
|
|||
|
|
- Sample Rate: 48kHz
|
|||
|
|
- Channels: Mono
|
|||
|
|
- Byte order: Native (little-endian on x86)
|
|||
|
|
|
|||
|
|
**Error (JSON):**
|
|||
|
|
```json
|
|||
|
|
{
|
|||
|
|
"error": "Error message"
|
|||
|
|
}
|
|||
|
|
```
|
|||
|
|
|
|||
|
|
## Behavior
|
|||
|
|
|
|||
|
|
### Automatic Synthesis Triggers
|
|||
|
|
|
|||
|
|
The server will automatically synthesize and send audio when:
|
|||
|
|
|
|||
|
|
1. **Sentence boundary**: `.` `!` `?` `。` `!` `?`
|
|||
|
|
2. **Pause boundary**: `,` `;` `,` `、`
|
|||
|
|
3. **Buffer limit**: More than 200 characters accumulated
|
|||
|
|
4. **Explicit flush**: Client sends `{"flush": true}`
|
|||
|
|
|
|||
|
|
### Example Flow
|
|||
|
|
|
|||
|
|
```
|
|||
|
|
Client: {"token": "Hello"}
|
|||
|
|
Client: {"token": " "}
|
|||
|
|
Client: {"token": "world"}
|
|||
|
|
Client: {"token": "!"}
|
|||
|
|
↓ Server detects '!' (sentence boundary)
|
|||
|
|
Server: [binary audio data for "Hello world!"]
|
|||
|
|
|
|||
|
|
Client: {"token": " "}
|
|||
|
|
Client: {"token": "How"}
|
|||
|
|
Client: {"token": " "}
|
|||
|
|
Client: {"token": "are"}
|
|||
|
|
Client: {"token": " "}
|
|||
|
|
Client: {"token": "you"}
|
|||
|
|
Client: {"token": "?"}
|
|||
|
|
↓ Server detects '?' (sentence boundary)
|
|||
|
|
Server: [binary audio data for "How are you?"]
|
|||
|
|
```
|
|||
|
|
|
|||
|
|
## Integration Examples
|
|||
|
|
|
|||
|
|
### Python with websockets
|
|||
|
|
|
|||
|
|
```python
|
|||
|
|
import websockets
|
|||
|
|
import json
|
|||
|
|
import numpy as np
|
|||
|
|
|
|||
|
|
async def speak(text_stream):
|
|||
|
|
async with websockets.connect('ws://localhost:8765/ws/stream') as ws:
|
|||
|
|
async for token in text_stream:
|
|||
|
|
# Send token
|
|||
|
|
await ws.send(json.dumps({"token": token, "pitch_shift": 0}))
|
|||
|
|
|
|||
|
|
# Receive audio (non-blocking)
|
|||
|
|
try:
|
|||
|
|
audio_bytes = await asyncio.wait_for(ws.recv(), timeout=0.5)
|
|||
|
|
audio = np.frombuffer(audio_bytes, dtype=np.float32)
|
|||
|
|
# Play audio...
|
|||
|
|
except asyncio.TimeoutError:
|
|||
|
|
continue # No audio yet
|
|||
|
|
|
|||
|
|
# Flush remaining
|
|||
|
|
await ws.send(json.dumps({"flush": True}))
|
|||
|
|
while True:
|
|||
|
|
try:
|
|||
|
|
audio_bytes = await asyncio.wait_for(ws.recv(), timeout=1.0)
|
|||
|
|
# Process final chunks...
|
|||
|
|
except asyncio.TimeoutError:
|
|||
|
|
break
|
|||
|
|
```
|
|||
|
|
|
|||
|
|
### Discord.py Integration
|
|||
|
|
|
|||
|
|
```python
|
|||
|
|
import discord
|
|||
|
|
from discord.ext import commands
|
|||
|
|
import websockets
|
|||
|
|
import json
|
|||
|
|
import io
|
|||
|
|
import asyncio
|
|||
|
|
|
|||
|
|
class MikuVoice(commands.Cog):
|
|||
|
|
def __init__(self, bot):
|
|||
|
|
self.bot = bot
|
|||
|
|
self.ws_url = 'ws://localhost:8765/ws/stream'
|
|||
|
|
|
|||
|
|
async def stream_to_discord(self, voice_client, text_stream):
|
|||
|
|
"""
|
|||
|
|
Stream TTS audio to Discord voice channel as LLM tokens arrive.
|
|||
|
|
|
|||
|
|
Args:
|
|||
|
|
voice_client: discord.VoiceClient
|
|||
|
|
text_stream: Async generator yielding text tokens
|
|||
|
|
"""
|
|||
|
|
async with websockets.connect(self.ws_url) as ws:
|
|||
|
|
# Queue for audio chunks
|
|||
|
|
audio_queue = asyncio.Queue()
|
|||
|
|
|
|||
|
|
# Task to play audio from queue
|
|||
|
|
async def player():
|
|||
|
|
while True:
|
|||
|
|
audio_bytes = await audio_queue.get()
|
|||
|
|
if audio_bytes is None: # Sentinel
|
|||
|
|
break
|
|||
|
|
|
|||
|
|
# Create Discord audio source
|
|||
|
|
audio_source = discord.FFmpegPCMAudio(
|
|||
|
|
io.BytesIO(audio_bytes),
|
|||
|
|
pipe=True,
|
|||
|
|
options='-f f32le -ar 48000 -ac 1'
|
|||
|
|
)
|
|||
|
|
|
|||
|
|
# Play (wait for previous to finish)
|
|||
|
|
while voice_client.is_playing():
|
|||
|
|
await asyncio.sleep(0.1)
|
|||
|
|
|
|||
|
|
voice_client.play(audio_source)
|
|||
|
|
|
|||
|
|
# Start player task
|
|||
|
|
player_task = asyncio.create_task(player())
|
|||
|
|
|
|||
|
|
# Stream tokens
|
|||
|
|
try:
|
|||
|
|
async for token in text_stream:
|
|||
|
|
# Send token to TTS
|
|||
|
|
await ws.send(json.dumps({
|
|||
|
|
"token": token,
|
|||
|
|
"pitch_shift": 0
|
|||
|
|
}))
|
|||
|
|
|
|||
|
|
# Receive audio (non-blocking)
|
|||
|
|
try:
|
|||
|
|
audio_bytes = await asyncio.wait_for(
|
|||
|
|
ws.recv(),
|
|||
|
|
timeout=0.5
|
|||
|
|
)
|
|||
|
|
await audio_queue.put(audio_bytes)
|
|||
|
|
except asyncio.TimeoutError:
|
|||
|
|
continue
|
|||
|
|
|
|||
|
|
# Flush remaining buffer
|
|||
|
|
await ws.send(json.dumps({"flush": True}))
|
|||
|
|
|
|||
|
|
# Get remaining audio
|
|||
|
|
while True:
|
|||
|
|
try:
|
|||
|
|
audio_bytes = await asyncio.wait_for(
|
|||
|
|
ws.recv(),
|
|||
|
|
timeout=1.0
|
|||
|
|
)
|
|||
|
|
await audio_queue.put(audio_bytes)
|
|||
|
|
except asyncio.TimeoutError:
|
|||
|
|
break
|
|||
|
|
|
|||
|
|
finally:
|
|||
|
|
# Signal player to stop
|
|||
|
|
await audio_queue.put(None)
|
|||
|
|
await player_task
|
|||
|
|
|
|||
|
|
@commands.command()
|
|||
|
|
async def speak(self, ctx, *, prompt: str):
|
|||
|
|
"""Make Miku speak in voice channel with streaming TTS"""
|
|||
|
|
|
|||
|
|
# Connect to voice if needed
|
|||
|
|
if not ctx.voice_client:
|
|||
|
|
if not ctx.author.voice:
|
|||
|
|
await ctx.send("You need to be in a voice channel!")
|
|||
|
|
return
|
|||
|
|
await ctx.author.voice.channel.connect()
|
|||
|
|
|
|||
|
|
# Get LLM response stream (example with llamacpp)
|
|||
|
|
async def llm_stream():
|
|||
|
|
# Replace with your actual LLM streaming code
|
|||
|
|
response = await your_llm_client.stream(prompt)
|
|||
|
|
async for token in response:
|
|||
|
|
yield token
|
|||
|
|
|
|||
|
|
# Stream to Discord
|
|||
|
|
await self.stream_to_discord(ctx.voice_client, llm_stream())
|
|||
|
|
await ctx.send("✓ Done speaking!")
|
|||
|
|
|
|||
|
|
async def setup(bot):
|
|||
|
|
await bot.add_cog(MikuVoice(bot))
|
|||
|
|
```
|
|||
|
|
|
|||
|
|
### JavaScript/Node.js
|
|||
|
|
|
|||
|
|
```javascript
|
|||
|
|
const WebSocket = require('ws');
|
|||
|
|
|
|||
|
|
async function streamTTS(tokens) {
|
|||
|
|
const ws = new WebSocket('ws://localhost:8765/ws/stream');
|
|||
|
|
|
|||
|
|
ws.on('open', () => {
|
|||
|
|
// Send tokens
|
|||
|
|
for (const token of tokens) {
|
|||
|
|
ws.send(JSON.stringify({
|
|||
|
|
token: token,
|
|||
|
|
pitch_shift: 0
|
|||
|
|
}));
|
|||
|
|
}
|
|||
|
|
|
|||
|
|
// Flush
|
|||
|
|
ws.send(JSON.stringify({ flush: true }));
|
|||
|
|
});
|
|||
|
|
|
|||
|
|
ws.on('message', (data) => {
|
|||
|
|
// data is Buffer containing PCM float32 audio
|
|||
|
|
const samples = new Float32Array(
|
|||
|
|
data.buffer,
|
|||
|
|
data.byteOffset,
|
|||
|
|
data.length / 4
|
|||
|
|
);
|
|||
|
|
|
|||
|
|
// Play audio...
|
|||
|
|
playAudio(samples);
|
|||
|
|
});
|
|||
|
|
}
|
|||
|
|
```
|
|||
|
|
|
|||
|
|
## Performance Characteristics
|
|||
|
|
|
|||
|
|
### Latency Breakdown
|
|||
|
|
|
|||
|
|
**Token-by-token (recommended):**
|
|||
|
|
```
|
|||
|
|
LLM token → Bot (5ms) → WebSocket (5ms) → Soprano (80ms) → RVC (100ms) → Discord (20ms)
|
|||
|
|
Total: ~210ms from token to sound
|
|||
|
|
```
|
|||
|
|
|
|||
|
|
**Sentence-by-sentence:**
|
|||
|
|
```
|
|||
|
|
Full sentence (1000ms) → WebSocket (5ms) → Soprano (200ms) → RVC (300ms) → Discord (20ms)
|
|||
|
|
Total: ~1525ms from start to sound
|
|||
|
|
```
|
|||
|
|
|
|||
|
|
### Throughput
|
|||
|
|
|
|||
|
|
- **Audio generation**: ~0.95x realtime (GPU accelerated)
|
|||
|
|
- **Network overhead**: <1% (binary protocol)
|
|||
|
|
- **Concurrent connections**: 10+ supported
|
|||
|
|
|
|||
|
|
## Audio Format Details
|
|||
|
|
|
|||
|
|
### Raw PCM Format
|
|||
|
|
|
|||
|
|
The WebSocket sends raw PCM audio data:
|
|||
|
|
|
|||
|
|
```python
|
|||
|
|
# Receiving and converting to numpy
|
|||
|
|
audio_bytes = await websocket.recv()
|
|||
|
|
audio = np.frombuffer(audio_bytes, dtype=np.float32)
|
|||
|
|
|
|||
|
|
# Audio properties
|
|||
|
|
sample_rate = 48000 # Hz
|
|||
|
|
channels = 1 # Mono
|
|||
|
|
dtype = np.float32 # 32-bit float
|
|||
|
|
value_range = [-1.0, 1.0] # Normalized
|
|||
|
|
```
|
|||
|
|
|
|||
|
|
### Converting to Other Formats
|
|||
|
|
|
|||
|
|
**To WAV:**
|
|||
|
|
```python
|
|||
|
|
import wave
|
|||
|
|
import struct
|
|||
|
|
|
|||
|
|
with wave.open('output.wav', 'wb') as wav:
|
|||
|
|
wav.setnchannels(1)
|
|||
|
|
wav.setsampwidth(4) # 4 bytes = float32
|
|||
|
|
wav.setframerate(48000)
|
|||
|
|
wav.writeframes(audio_bytes)
|
|||
|
|
```
|
|||
|
|
|
|||
|
|
**To Discord Opus:**
|
|||
|
|
```python
|
|||
|
|
import discord
|
|||
|
|
|
|||
|
|
# Discord expects PCM s16le (16-bit signed integer)
|
|||
|
|
audio_float = np.frombuffer(audio_bytes, dtype=np.float32)
|
|||
|
|
audio_int16 = (audio_float * 32767).astype(np.int16)
|
|||
|
|
audio_source = discord.PCMAudio(io.BytesIO(audio_int16.tobytes()))
|
|||
|
|
```
|
|||
|
|
|
|||
|
|
## Best Practices
|
|||
|
|
|
|||
|
|
### 1. Token Buffering
|
|||
|
|
|
|||
|
|
Don't send every character individually - send word-by-word or phrase-by-phrase:
|
|||
|
|
|
|||
|
|
```python
|
|||
|
|
# ✗ Bad: Too granular
|
|||
|
|
for char in text:
|
|||
|
|
await ws.send(json.dumps({"token": char}))
|
|||
|
|
|
|||
|
|
# ✓ Good: Word-by-word
|
|||
|
|
for word in text.split():
|
|||
|
|
await ws.send(json.dumps({"token": " " + word}))
|
|||
|
|
```
|
|||
|
|
|
|||
|
|
### 2. Error Handling
|
|||
|
|
|
|||
|
|
Always handle disconnections gracefully:
|
|||
|
|
|
|||
|
|
```python
|
|||
|
|
try:
|
|||
|
|
async with websockets.connect(url) as ws:
|
|||
|
|
# ... streaming code ...
|
|||
|
|
except websockets.exceptions.ConnectionClosed:
|
|||
|
|
logger.error("Connection lost, reconnecting...")
|
|||
|
|
# Retry logic...
|
|||
|
|
```
|
|||
|
|
|
|||
|
|
### 3. Backpressure
|
|||
|
|
|
|||
|
|
If Discord's audio buffer fills, slow down token sending:
|
|||
|
|
|
|||
|
|
```python
|
|||
|
|
if voice_client.is_playing():
|
|||
|
|
await asyncio.sleep(0.1) # Wait for buffer space
|
|||
|
|
```
|
|||
|
|
|
|||
|
|
### 4. Flush at End
|
|||
|
|
|
|||
|
|
Always flush to ensure all audio is sent:
|
|||
|
|
|
|||
|
|
```python
|
|||
|
|
# After sending all tokens
|
|||
|
|
await ws.send(json.dumps({"flush": True}))
|
|||
|
|
|
|||
|
|
# Wait for remaining audio
|
|||
|
|
try:
|
|||
|
|
while True:
|
|||
|
|
await asyncio.wait_for(ws.recv(), timeout=1.0)
|
|||
|
|
except asyncio.TimeoutError:
|
|||
|
|
pass # All audio received
|
|||
|
|
```
|
|||
|
|
|
|||
|
|
## Troubleshooting
|
|||
|
|
|
|||
|
|
### No Audio Received
|
|||
|
|
|
|||
|
|
**Problem**: Sending tokens but no audio comes back
|
|||
|
|
|
|||
|
|
**Solutions**:
|
|||
|
|
1. Check if you're hitting sentence boundaries (`.` `!` `?`)
|
|||
|
|
2. Try sending `{"flush": true}` manually
|
|||
|
|
3. Verify token format: `{"token": "text", "pitch_shift": 0}`
|
|||
|
|
|
|||
|
|
### Audio Choppy/Gaps
|
|||
|
|
|
|||
|
|
**Problem**: Audio plays but with interruptions
|
|||
|
|
|
|||
|
|
**Solutions**:
|
|||
|
|
1. Increase buffer size on Discord side
|
|||
|
|
2. Send tokens in larger chunks (word-by-word, not char-by-char)
|
|||
|
|
3. Check network latency: `ping localhost`
|
|||
|
|
|
|||
|
|
### Connection Drops
|
|||
|
|
|
|||
|
|
**Problem**: WebSocket disconnects unexpectedly
|
|||
|
|
|
|||
|
|
**Solutions**:
|
|||
|
|
1. Implement reconnection logic with exponential backoff
|
|||
|
|
2. Send periodic ping/pong frames
|
|||
|
|
3. Check Docker container logs: `docker logs miku-rvc-api`
|
|||
|
|
|
|||
|
|
### High Latency
|
|||
|
|
|
|||
|
|
**Problem**: Long delay between token and audio
|
|||
|
|
|
|||
|
|
**Solutions**:
|
|||
|
|
1. Verify GPU is being used: Check logs for "Found GPU AMD Radeon RX 6800"
|
|||
|
|
2. Reduce sentence buffer triggers (adjust in code)
|
|||
|
|
3. Use smaller chunks: `chunk_size=5` in Soprano config
|
|||
|
|
|
|||
|
|
## Comparison with HTTP API
|
|||
|
|
|
|||
|
|
| Feature | WebSocket (`/ws/stream`) | HTTP (`/api/speak`) |
|
|||
|
|
|---------|-------------------------|-------------------|
|
|||
|
|
| Latency | ~200ms | ~1500ms |
|
|||
|
|
| Streaming | ✓ Token-by-token | ✗ Request-response |
|
|||
|
|
| Overhead | 5ms per message | 100ms per request |
|
|||
|
|
| Connection | Persistent | Per-request |
|
|||
|
|
| Backpressure | ✓ Bidirectional | ✗ One-way |
|
|||
|
|
| Complexity | Medium | Low |
|
|||
|
|
| Use Case | Real-time voice chat | Simple TTS requests |
|
|||
|
|
|
|||
|
|
**Recommendation**: Use WebSocket for Discord bot, keep HTTP for testing/debugging.
|