Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

BA-MoE: Boundary-Aware Mixture-of-Experts Adapter for Code-Switching Speech Recognition (2310.02629v2)

Published 4 Oct 2023 in cs.SD and eess.AS

Abstract: Mixture-of-experts based models, which use language experts to extract language-specific representations effectively, have been well applied in code-switching automatic speech recognition. However, there is still substantial space to improve as similar pronunciation across languages may result in ineffective multi-LLMing and inaccurate language boundary estimation. To eliminate these drawbacks, we propose a cross-layer language adapter and a boundary-aware training method, namely Boundary-Aware Mixture-of-Experts (BA-MoE). Specifically, we introduce language-specific adapters to separate language-specific representations and a unified gating layer to fuse representations within each encoder layer. Second, we compute language adaptation loss of the mean output of each language-specific adapter to improve the adapter module's language-specific representation learning. Besides, we utilize a boundary-aware predictor to learn boundary representations for dealing with language boundary confusion. Our approach achieves significant performance improvement, reducing the mixture error rate by 16.55\% compared to the baseline on the ASRU 2019 Mandarin-English code-switching challenge dataset.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Peikun Chen (9 papers)
  2. Fan Yu (63 papers)
  3. Yuhao Lian (2 papers)
  4. Hongfei Xue (22 papers)
  5. Xucheng Wan (12 papers)
  6. Naijun Zheng (8 papers)
  7. Huan Zhou (51 papers)
  8. Lei Xie (337 papers)
Citations (6)