Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Internal Language Model Estimation based Adaptive Language Model Fusion for Domain Adaptation (2211.00968v2)

Published 2 Nov 2022 in cs.CL, cs.SD, and eess.AS

Abstract: ASR model deployment environment is ever-changing, and the incoming speech can be switched across different domains during a session. This brings a challenge for effective domain adaptation when only target domain text data is available, and our objective is to obtain obviously improved performance on the target domain while the performance on the general domain is less undermined. In this paper, we propose an adaptive LM fusion approach called internal LLM estimation based adaptive domain adaptation (ILME-ADA). To realize such an ILME-ADA, an interpolated log-likelihood score is calculated based on the maximum of the scores from the internal LM and the external LM (ELM) respectively. We demonstrate the efficacy of the proposed ILME-ADA method with both RNN-T and LAS modeling frameworks employing neural network and n-gram LMs as ELMs respectively on two domain specific (target) test sets. The proposed method can achieve significantly better performance on the target test sets while it gets minimal performance degradation on the general test set, compared with both shallow and ILME-based LM fusion methods.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Rao Ma (22 papers)
  2. Xiaobo Wu (3 papers)
  3. Jin Qiu (5 papers)
  4. Yanan Qin (1 paper)
  5. Haihua Xu (23 papers)
  6. Peihao Wu (8 papers)
  7. Zejun Ma (78 papers)
Citations (2)