Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Meta-Transfer Learning for Code-Switched Speech Recognition (2004.14228v1)

Published 29 Apr 2020 in cs.CL, cs.SD, and eess.AS

Abstract: An increasing number of people in the world today speak a mixed-language as a result of being multilingual. However, building a speech recognition system for code-switching remains difficult due to the availability of limited resources and the expense and significant effort required to collect mixed-language data. We therefore propose a new learning method, meta-transfer learning, to transfer learn on a code-switched speech recognition system in a low-resource setting by judiciously extracting information from high-resource monolingual datasets. Our model learns to recognize individual languages, and transfer them so as to better recognize mixed-language speech by conditioning the optimization on the code-switching data. Based on experimental results, our model outperforms existing baselines on speech recognition and LLMing tasks, and is faster to converge.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Genta Indra Winata (94 papers)
  2. Samuel Cahyawijaya (75 papers)
  3. Zhaojiang Lin (45 papers)
  4. Zihan Liu (102 papers)
  5. Peng Xu (357 papers)
  6. Pascale Fung (151 papers)
Citations (53)