Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Proteno: Text Normalization with Limited Data for Fast Deployment in Text to Speech Systems (2104.07777v1)

Published 15 Apr 2021 in cs.CL

Abstract: Developing Text Normalization (TN) systems for Text-to-Speech (TTS) on new languages is hard. We propose a novel architecture to facilitate it for multiple languages while using data less than 3% of the size of the data used by the state of the art results on English. We treat TN as a sequence classification problem and propose a granular tokenization mechanism that enables the system to learn majority of the classes and their normalizations from the training data itself. This is further combined with minimal precoded linguistic knowledge for other classes. We publish the first results on TN for TTS in Spanish and Tamil and also demonstrate that the performance of the approach is comparable with the previous work done on English. All annotated datasets used for experimentation will be released at https://github.com/amazon-research/proteno.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Shubhi Tyagi (5 papers)
  2. Antonio Bonafonte (14 papers)
  3. Jaime Lorenzo-Trueba (33 papers)
  4. Javier Latorre (6 papers)
Citations (12)

Summary

We haven't generated a summary for this paper yet.