Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Prompt Tuning GPT-2 language model for parameter-efficient domain adaptation of ASR systems (2112.08718v3)

Published 16 Dec 2021 in cs.CL and cs.LG

Abstract: Automatic Speech Recognition (ASR) systems have found their use in numerous industrial applications in very diverse domains creating a need to adapt to new domains with small memory and deployment overhead. In this work, we introduce domain-prompts, a methodology that involves training a small number of domain embedding parameters to prime a Transformer-based LLM (LM) to a particular domain. Using this domain-adapted LM for rescoring ASR hypotheses can achieve 7-13% WER reduction for a new domain with just 1000 unlabeled textual domain-specific sentences. This improvement is comparable or even better than fully fine-tuned models even though just 0.02% of the parameters of the base LM are updated. Additionally, our method is deployment-friendly as the learnt domain embeddings are prefixed to the input to the model rather than changing the base model architecture. Therefore, our method is an ideal choice for on-the-fly adaptation of LMs used in ASR systems to progressively scale it to new domains.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Saket Dingliwal (22 papers)
  2. Ashish Shenoy (13 papers)
  3. Sravan Bodapati (31 papers)
  4. Ankur Gandhe (30 papers)
  5. Ravi Teja Gadde (6 papers)
  6. Katrin Kirchhoff (36 papers)
Citations (3)