OZSpeech: Zero-Shot TTS with One-Step Inference
- OZSpeech is a zero-shot text-to-speech synthesis framework that leverages optimal-transport conditional flow matching and factorized speech token representations for robust one-step inference.
- It decomposes speech into six discrete token streams from a pre-trained neural codec, enabling precise control over content, prosody, acoustics, and timbre.
- The system achieves real-time synthesis with reduced computation, outperforming traditional autoregressive and multi-step methods in efficiency and speaker cloning.
OZSpeech is a zero-shot text-to-speech (TTS) synthesis framework that advances efficient, high-quality, and speaker-adaptive speech generation by integrating a learned discrete prior, factorized speech token representations, and an optimal-transport conditional flow-matching (OT-CFM) mechanism to enable single-step inference. Unlike autoregressive codec–LLMs and multi-step diffusion or flow-matching methods, OZSpeech achieves robust zero-shot speaker cloning and accurate speech synthesis with significantly reduced compute and sampling cost by leveraging a factorized neural codec and a novel application of OT-CFM with a learned prior (Huynh-Nguyen et al., 19 May 2025).
1. Motivation and Model Overview
OZSpeech addresses principal zero-shot TTS challenges, specifically speaker generalization from minimal prompts, robust disentanglement of speech attributes (content, prosody, acoustic detail, timbre), and sampling efficiency. Conventional models—either autoregressive codec-LLMs (e.g., VALL-E) or multi-step flow-matching systems (e.g., E2 TTS, F5-TTS)—suffer from inefficiencies due to hundreds of iterative sampling steps and entangled attribute modeling, limiting real-time utility and speaker/style fidelity.
OZSpeech decomposes speech into six discrete token streams (from a pre-trained factorized neural codec "FACodec": two content, one prosody, three acoustic), and a timbre embedding. A neural prior generator predicts intermediate token codes from phoneme sequences, positioned to approximate the FACodec codes of target speech. A conditional vector field estimator then refines these tokens to FACodec-aligned codes in a single step via an OT-CFM objective, where all previous state dependencies are removed, and the learned prior acts as the starting distribution.
2. Factorized Codec Tokenization and Representation
OZSpeech utilizes FACodec, a pre-trained codec-based speech tokenization framework. FACodec maps input waveforms to an encoder output , which is then quantized via three independently trained factorized vector quantizers (FVQs):
- Prosody: sequence,
- Content: sequences,
- Acoustic details: sequences,
Concatenation yields , with discrete codes in . Timbre is modeled by . Quantizer track identifiers are added, and a folding operation projects , permitting efficient Transformer-based modeling of joint code streams.
3. Learned-Prior Conditional Flow Matching Framework
The TTS pipeline components are:
Prior Codes Generator
The prior generator models the joint distribution over quantizer streams:
with associated loss . A duration-predictor loss aligns phonemes to code length.
Conditional OT-CFM
Classical conditional flow matching considers , , and linearly mixes states via , . The OT-CFM target is:
OZSpeech replaces the Gaussian start with the learned prior , optimizing an implicit per sample:
An anchor loss is used for discrete codes:
with the total loss:
4. One-Step Inference and System Efficiency
The one-step generative process consists of:
- Generate prior codes .
- Apply folding and quantizer encoding to the concatenated prompt-masked and prior codes.
- Feed codes and implicit time to the vector field estimator .
- Refine:
- Decode to waveform with FACodec.
Numerical Function Evaluations (NFE) are reduced to one, enabling real-time synthesis. Real-time factor (RTF) is 0.26 speech-sec/sec, compared to 0.7–1.7 for prior art, with only 145M trainable parameters plus 102M for FACodec.
5. Speaker Cloning, Robustness, and Evaluation
OZSpeech exploits "acoustic prompt injection" by masking content codes within the prompt, conditioning generation on prosody and acoustic tokens for speaker style preservation. The learned prior aligns generated codes with text duration, allowing the vector field to apply speaker-specific pitch and timbre adaptation. Word Error Rate (WER) remains stable (~0.05) down to 0 dB SNR, substantially surpassing baseline noise robustness.
Key empirical results with 3 s prompt length:
| Metric | OZSpeech | Best Baseline Range |
|---|---|---|
| WER | 0.05 | 0.09–0.24 |
| UTMOS | 3.15 | 3.55–3.76 |
| SIM-O/SIM-R | 0.40/0.47 | 0.31–0.53/0.38–0.51 |
| F0 Accuracy | 0.81 | ≤0.69 |
| RMSE (F0) | 11.96 | ≥12.96 |
| NFE | 1 | 32–200 |
| RTF (sec/sec) | 0.26 | 0.7–1.7 |
| Model size (M) | 145 | 378–830 |
Ablation studies demonstrate that arbitrary segment prompting outperforms using only the first segment. OZSpeech-Small (100M parameters) matches the performance of the Base model (145M) on WER and UTMOS, with a minimal tradeoff in prosody.
6. Limitations and Prospects
OZSpeech exhibits slight temporal distortions from rounding in the FastSpeech duration predictor, and its UTMOS naturalness score trails some larger models. Prospective improvements include alternative alignment strategies (e.g., Monotonic Alignment Search) to enhance temporal fidelity and adaptive noise filtering. Multilingual and multimodal zero-shot synthesis constitute additional future directions.
7. Comparative Context and Significance
OZSpeech constitutes the first application of optimal-transport conditional flow matching with a learned discrete prior and single-step sampling in TTS, using a fully disentangled, factorized token representation. This approach departs from previous zero-resource and zero-shot paradigms by both eliminating multi-step sampling and enabling fine-grained attribute control within a compact architecture. The combination of state-of-the-art content accuracy (WER ≈ 0.05), robust prosody/style transfer, resistance to noise, and high sampling efficiency establishes a new practical benchmark for zero-shot TTS, with implications for scalable, real-time, speaker-adaptive speech synthesis (Huynh-Nguyen et al., 19 May 2025).
Sponsored by Paperpile, the PDF & BibTeX manager trusted by top AI labs.
Get 30 days free