2000 character limit reached
Effect and Analysis of Large-scale Language Model Rescoring on Competitive ASR Systems (2204.00212v2)
Published 1 Apr 2022 in cs.CL, cs.SD, and eess.AS
Abstract: Large-scale LLMs such as GPT-2, BERT and RoBERTa have been successfully applied to ASR N-best rescoring. However, whether or how they can benefit competitive, near state-of-the-art ASR systems remains unexplored. In this study, we incorporate LLM rescoring into one of the most competitive ASR baselines: the Conformer-Transducer model. We demonstrate that consistent improvement is achieved by the LLM's bidirectionality, pretraining, in-domain finetuning and context augmentation. Furthermore, our lexical analysis sheds light on how each of these components may be contributing to the ASR performance.
- Takuma Udagawa (18 papers)
- Masayuki Suzuki (6 papers)
- Gakuto Kurata (13 papers)
- Nobuyasu Itoh (1 paper)
- George Saon (39 papers)