Runtime Text To Speech

Cross-platform, offline text-to-speech synthesis for Unreal Engine. Generate natural-sounding speech with 900+ voices across 47 languages - including Kokoro studio-quality voice models and streaming TTS - with full Blueprint and C++ support.

UE 4.27 – 5.7
Blueprints & C++
All platforms supported
47 languages, 900+ voices

How It Works

Text is synthesized entirely on-device using Piper or Kokoro voice models via ONNX Runtime - standard or streaming - with no internet connection, no API keys, and no data sent externally. The resulting PCM audio integrates directly with Runtime Audio Importer for immediate playback.

1

Text Input

Any string from Blueprint, C++, or piped in from an AI chatbot response

2

Voice Model

Select a Piper or Kokoro model by name or object, with optional speaker and quality settings

3

On-Device Synthesis

ONNX Runtime processes the text locally - standard or streaming - with no data sent externally

4

PCM Audio Output

Float32 PCM data ready for playback, saving, lip sync, or any downstream audio processing

Voice Model Types

Two model architectures covering a wide range of use cases, from broad language coverage to studio-quality output.

Piper Voice Models

ONNX-based models covering 47 languages with 303 voice models and 170+ quality variants. Supports multi-speaker models with up to 900+ speakers. Accepts custom-trained ONNX models.

Kokoro Studio-Quality Models

151 high-quality open-source voice models across 8 languages: English (US & UK), Simplified Chinese, Spanish, Portuguese, Hindi, French, and Italian. Among the highest-quality open-source TTS solutions available.

Custom Voice Models

Import your own Piper (ONNX + JSON) or Kokoro (BIN + JSON) voice models directly via the plugin settings for specialized voices or languages.

Languages & Platform Support

Broad language coverage and first-class support for every major platform Unreal Engine targets.

Language Coverage

English (US & UK)
German
Spanish
French
Simplified Chinese
Russian
Portuguese
Hindi
Italian
Polish
37 more - Dutch, Turkish, Korean, Vietnamese, Ukrainian, Catalan...

Platform Support

Windows
Mac
Linux
Android
iOS
Meta Quest

Full-Featured Speech Control

Beyond basic synthesis, the plugin provides streaming output, multi-speaker selection, cancellable operations, and model management utilities - all accessible from Blueprints or C++.

Standard & Streaming Synthesis

Two synthesis modes: complete the full text before returning audio, or stream PCM chunks in real-time as they're generated for immediate playback of long texts.

Multi-Speaker Selection

Many models support multiple speakers - English LibriTTS includes 900+ distinct voices. Query speaker count per model and select by index at synthesis time.

Cancellable Operations

Cancel any ongoing synthesis at any time - essential for responsive applications that need to interrupt and restart speech mid-generation.

In-Editor Model Management

Download, preview, and manage voice models directly from the Unreal Editor. List available models, query metadata, and control storage without leaving the engine.

Voice Metadata & Utilities

Retrieve model metadata, language, quality tier, and speaker count by name or object reference. Select models dynamically at runtime based on language or other criteria.

Fully On-Device

No internet connection, no API keys, no data leaving the device. Suitable for privacy-sensitive applications, air-gapped environments, and offline games.

Blueprint Example - Streaming TTS with real-time playback

Streaming TTS with Runtime Audio Importer blueprint example

Try the Demo

A Windows demo is available to experience the plugin's voice quality firsthand. The demo includes a curated selection of Piper and Kokoro voice models showcasing different languages, speaker counts, and synthesis modes.

Multiple voice model types

Standard, multi-speaker, and Kokoro studio-quality models

Standard and streaming synthesis

Compare both synthesis modes side by side with custom text input

Blueprint source included

Full Blueprint implementation in the demo project for reference

Download Demo (Windows)

Video Tutorial

Plugin Ecosystem

Runtime Text To Speech is the voice output layer of the Georgy Dev plugin suite - driving lip sync, powering AI character pipelines, and completing the speech recognition loop.

Runtime Audio Importer

Receive PCM audio output from TTS synthesis and play it back, stream it, or process it further at runtime.

Learn more

Runtime MetaHuman Lip Sync

Drive real-time lip sync animations on MetaHuman and custom characters directly from TTS audio output.

Learn more

Runtime Speech Recognizer

Close the voice loop - pair offline speech recognition with offline TTS for fully on-device conversational experiences.

Learn more

AI Chatbot Integrator

Pipe AI-generated text from OpenAI, Claude, or DeepSeek directly into TTS synthesis for fully voiced NPC responses.

Learn more

Documentation & Support

Comprehensive documentation covers all synthesis modes, model management, multi-speaker configuration, streaming workflows, and platform-specific guidance.

Full Documentation

Step-by-step guides for all features, with Blueprint and C++ examples

Community & Support

Active Discord community with developer support

Custom Development

Tailored integration or feature development - [email protected]

Demo - Voice model download and preview in the editor

Download and preview voice models in the editor

Ready to Add Text-to-Speech to Your Project?

Available on Fab for UE 4.27 – 5.7. Includes all voice models, full documentation, and a demo project showcasing streaming TTS and Kokoro voices.