Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Prompt-tuning in ASR systems for efficient domain-adaptation (2110.06502v2)

Published 13 Oct 2021 in cs.CL

Abstract: Automatic Speech Recognition (ASR) systems have found their use in numerous industrial applications in very diverse domains. Since domain-specific systems perform better than their generic counterparts on in-domain evaluation, the need for memory and compute-efficient domain adaptation is obvious. Particularly, adapting parameter-heavy transformer-based LLMs used for rescoring ASR hypothesis is challenging. In this work, we overcome the problem using prompt-tuning, a methodology that trains a small number of domain token embedding parameters to prime a transformer-based LM to a particular domain. With just a handful of extra parameters per domain, we achieve much better perplexity scores over the baseline of using an unadapted LM. Despite being parameter-efficient, these improvements are comparable to those of fully-fine-tuned models with hundreds of millions of parameters. We replicate our findings in perplexity numbers to Word Error Rate in a domain-specific ASR system for one such domain.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Saket Dingliwal (22 papers)
  2. Ashish Shenoy (13 papers)
  3. Sravan Bodapati (31 papers)
  4. Ankur Gandhe (30 papers)
  5. Ravi Teja Gadde (6 papers)
  6. Katrin Kirchhoff (36 papers)
Citations (3)

Summary

We haven't generated a summary for this paper yet.