Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Unsupervised Data Selection via Discrete Speech Representation for ASR (2204.01981v1)

Published 5 Apr 2022 in eess.AS

Abstract: Self-supervised learning of speech representations has achieved impressive results in improving automatic speech recognition (ASR). In this paper, we show that data selection is important for self-supervised learning. We propose a simple and effective unsupervised data selection method which selects acoustically similar speech to a target domain. It takes the discrete speech representation available in common self-supervised learning frameworks as input, and applies a contrastive data selection method on the discrete tokens. Through extensive empirical studies we show that our proposed method reduces the amount of required pre-training data and improves the downstream ASR performance. Pre-training on a selected subset of 6% of the general data pool results in 11.8% relative improvements in LibriSpeech test-other compared to pre-training on the full set. On Multilingual LibriSpeech French, German, and Spanish test sets, selecting 6% data for pre-training reduces word error rate by more than 15% relatively compared to the full set, and achieves competitive results compared to current state-of-the-art performances.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Zhiyun Lu (19 papers)
  2. Yongqiang Wang (92 papers)
  3. Yu Zhang (1400 papers)
  4. Wei Han (202 papers)
  5. Zhehuai Chen (39 papers)
  6. Parisa Haghani (15 papers)
Citations (11)

Summary

We haven't generated a summary for this paper yet.