Orpheus TTS with LM Studio
Guide

Run Orpheus TTS Locally with LM Studio — Human-Like AI Speech

Feb 2025 · 8 min read · TTS · LM Studio · Orpheus · Local AI

What Is Orpheus TTS?

Orpheus is a speech language model released by Canopy AI under the Apache 2.0 license — meaning it's free to use, modify, and deploy for both personal and commercial projects. What sets it apart from other open-source TTS options is how genuinely human it sounds. It doesn't just read text flatly; it conveys emotion, varies intonation naturally, and supports expressive tags that let you control how speech is delivered.

🎭
Emotion Tags
Add tags like <laugh> or <sigh> to inject natural expressiveness into the generated voice.
🎤
8 Built-In Voices
Choose from a selection of distinct voices — each with its own character and tone.
Fast on RTX GPUs
Generates audio in under a minute on a 4060 6GB — optimized for NVIDIA RTX cards.
💻
Low VRAM
The GGUF version runs on GPUs with less than 4GB VRAM, accessible to most setups.
🌐
OpenAI-Compatible API
Exposes an API compatible with OpenAI's TTS endpoint — plug it into existing tools.
🔒
Fully Local
Everything runs on your machine. No audio sent to the cloud, no API costs, complete privacy.

The Setup: Orpheus FastAPI Web UI + LM Studio

To run Orpheus easily, we use the Orpheus FastAPI Web UI — an open-source project that wraps the model in a clean browser-based interface. The Local Lab fork of this project adds built-in LM Studio support out of the box, extends the context window to 8,192 tokens, and uses the GGUF model format for lower resource usage.

You'll need two things running together: LM Studio (which serves the Orpheus model via its local API) and the Orpheus FastAPI server (which provides the web UI and connects to LM Studio).

💡 Patreon one-click installer available If you'd rather skip the manual setup, a one-click Windows installer is available on The Local Lab Patreon. It handles all dependencies and includes startup files.

Step 1 — Set Up LM Studio and Load the Model

  1. Install LM Studio — download from lmstudio.ai and install for your OS (Windows, Mac, or Linux).
  2. Download the Orpheus Model — open LM Studio, go to the Discover tab, and search for Orpheus. Download the orpheus-3b-4k-gguf model — it's compact and runs smoothly on 4GB+ VRAM.
  3. Load the Model and Start the API Server — switch to the Developer tab in LM Studio, load the Orpheus model, and confirm the local API server starts on port 1234.

Step 2 — Install the Orpheus FastAPI Web UI (Manual)

We'll use Miniconda to keep the Python environment clean and isolated.

  1. Install Miniconda — download and install Miniconda from the Anaconda website. Once installed, open the Anaconda Prompt from your Windows search bar.
  2. Create and Activate a Conda Environment — create a dedicated Python 3.10 environment (required for compatibility):
conda create -n orpheus-tts python=3.10 -y conda activate orpheus-tts
  1. Clone the Repository — navigate to your preferred install folder, then clone The Local Lab's fork of the Orpheus FastAPI Web UI:
git clone https://github.com/TheLocalLab/Orpheus-FastAPI-LMStudio cd orpheus-fastapi-lm-studio
  1. Install PyTorch with CUDA and Dependencies — run the PyTorch install command from the repository README, then install project requirements:
pip install -r requirements.txt
  1. Launch the FastAPI Server — start with:
python app.py

Then open your browser and navigate to http://127.0.0.1:1555 to access the Orpheus TTS web UI.

Using the Web UI

Once both LM Studio (with Orpheus loaded) and the FastAPI server are running, the web UI is straightforward:

💡 Generation speed On an RTX 4060 6GB, short to medium text generates in under a minute. The Local Lab fork extends the context window to 8,192 tokens — significantly larger than the base project, so you can feed in long scripts without hitting a limit mid-generation.

📦 Want to skip the setup?

The Local Lab offers pre-configured AI installer packages so you can get running in minutes, not hours.

Get the Installer →