Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Dual-Pipeline with Low-Rank Adaptation for New Language Integration in Multilingual ASR (2406.07842v1)

Published 12 Jun 2024 in eess.AS and cs.CL

Abstract: This paper addresses challenges in integrating new languages into a pre-trained multilingual automatic speech recognition (mASR) system, particularly in scenarios where training data for existing languages is limited or unavailable. The proposed method employs a dual-pipeline with low-rank adaptation (LoRA). It maintains two data flow pipelines-one for existing languages and another for new languages. The primary pipeline follows the standard flow through the pre-trained parameters of mASR, while the secondary pipeline additionally utilizes language-specific parameters represented by LoRA and a separate output decoder module. Importantly, the proposed approach minimizes the performance degradation of existing languages and enables a language-agnostic operation mode, facilitated by a decoder selection strategy. We validate the effectiveness of the proposed method by extending the pre-trained Whisper model to 19 new languages from the FLEURS dataset

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Yerbolat Khassanov (19 papers)
  2. Zhipeng Chen (46 papers)
  3. Tianfeng Chen (3 papers)
  4. Tze Yuang Chong (3 papers)
  5. Wei Li (1122 papers)
  6. Jun Zhang (1008 papers)
  7. Lu Lu (189 papers)
  8. Yuxuan Wang (239 papers)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com