Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Magic dust for cross-lingual adaptation of monolingual wav2vec-2.0 (2110.03560v1)

Published 7 Oct 2021 in cs.CL, cs.SD, and eess.AS

Abstract: We propose a simple and effective cross-lingual transfer learning method to adapt monolingual wav2vec-2.0 models for Automatic Speech Recognition (ASR) in resource-scarce languages. We show that a monolingual wav2vec-2.0 is a good few-shot ASR learner in several languages. We improve its performance further via several iterations of Dropout Uncertainty-Driven Self-Training (DUST) by using a moderate-sized unlabeled speech dataset in the target language. A key finding of this work is that the adapted monolingual wav2vec-2.0 achieves similar performance as the topline multilingual XLSR model, which is trained on fifty-three languages, on the target language ASR task.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Sameer Khurana (26 papers)
  2. Antoine Laurent (22 papers)
  3. James Glass (173 papers)
Citations (18)

Summary

We haven't generated a summary for this paper yet.