Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Optimizing Byte-level Representation for End-to-end ASR (2406.09676v2)

Published 14 Jun 2024 in eess.AS and cs.CL

Abstract: We propose a novel approach to optimizing a byte-level representation for end-to-end automatic speech recognition (ASR). Byte-level representation is often used by large scale multilingual ASR systems when the character set of the supported languages is large. The compactness and universality of byte-level representation allow the ASR models to use smaller output vocabularies and therefore, provide more flexibility. UTF-8 is a commonly used byte-level representation for multilingual ASR, but it is not designed to optimize machine learning tasks directly. By using auto-encoder and vector quantization, we show that we can optimize a byte-level representation for ASR and achieve better accuracy. Our proposed framework can incorporate information from different modalities, and provides an error correction mechanism. In an English/Mandarin dictation task, we show that a bilingual ASR model built with this approach can outperform UTF-8 representation by 5% relative in error rate.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Roger Hsiao (10 papers)
  2. Liuhui Deng (4 papers)
  3. Erik McDermott (9 papers)
  4. Ruchir Travadi (6 papers)
  5. Xiaodan Zhuang (9 papers)
X Twitter Logo Streamline Icon: https://streamlinehq.com