Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Iterative Pseudo-Labeling for Speech Recognition (2005.09267v2)

Published 19 May 2020 in cs.CL, cs.SD, and eess.AS

Abstract: Pseudo-labeling has recently shown promise in end-to-end automatic speech recognition (ASR). We study Iterative Pseudo-Labeling (IPL), a semi-supervised algorithm which efficiently performs multiple iterations of pseudo-labeling on unlabeled data as the acoustic model evolves. In particular, IPL fine-tunes an existing model at each iteration using both labeled data and a subset of unlabeled data. We study the main components of IPL: decoding with a LLM and data augmentation. We then demonstrate the effectiveness of IPL by achieving state-of-the-art word-error rate on the Librispeech test sets in both standard and low-resource setting. We also study the effect of LLMs trained on different corpora to show IPL can effectively utilize additional text. Finally, we release a new large in-domain text corpus which does not overlap with the Librispeech training transcriptions to foster research in low-resource, semi-supervised ASR

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Qiantong Xu (26 papers)
  2. Tatiana Likhomanenko (41 papers)
  3. Jacob Kahn (21 papers)
  4. Awni Hannun (33 papers)
  5. Gabriel Synnaeve (97 papers)
  6. Ronan Collobert (55 papers)
Citations (128)

Summary

We haven't generated a summary for this paper yet.