Skip to content

Latest commit

 

History

History
24 lines (12 loc) · 1.08 KB

File metadata and controls

24 lines (12 loc) · 1.08 KB

Orpheus TTS on Baseten

Baseten is Canopy Labs' preferred inference provider for running Orpheus TTS in production.

Deployment

To deploy the model, go to https://www.baseten.co/library/orpheus-tts/ and use the one-click deploy option.

Baseten supports both fp8 (default for performance) and fp16 (full fidelity) versions of Orpheus.

If you want to customize the model serving code, you can instead deploy the prepackaged model from Baseten's example repository.

Inference

The call_orpheus.py file contains sample inference code for running the Orpheus TTS model with multiple parallel requests.

Prerequisites:

  • Paste the model_id from your deployed model into the call_orpheus.py script.
  • Save your BASETEN_API_KEY as an environment variable.

Then, you can call the model with python call_orpheus.py.