This page covers server-side inference for Fish Audio S2, plus quick links for WebUI inference and Docker deployment.
Fish Speech provides an HTTP API server entrypoint at tools/api_server.py.
python tools/api_server.py \
--llama-checkpoint-path checkpoints/s2-pro \
--decoder-checkpoint-path checkpoints/s2-pro/codec.pth \
--listen 0.0.0.0:8080
Common options:
--compile: enable torch.compile optimization--half: use fp16 mode--api-key: require bearer token authentication--workers: set worker process countcurl -X GET http://127.0.0.1:8080/v1/health
Expected response:
{"status":"ok"}
POST /v1/tts for text-to-speech generationPOST /v1/vqgan/encode for VQ encodePOST /v1/vqgan/decode for VQ decodeFor WebUI usage, see:
For Docker-based server or WebUI deployment, see:
You can also start the server profile directly with Docker Compose:
docker compose --profile server up