Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Don't Stop Self-Supervision: Accent Adaptation of Speech Representations via Residual Adapters (2307.00453v1)

Published 2 Jul 2023 in cs.CL, cs.SD, and eess.AS

Abstract: Speech representations learned in a self-supervised fashion from massive unlabeled speech corpora have been adapted successfully toward several downstream tasks. However, such representations may be skewed toward canonical data characteristics of such corpora and perform poorly on atypical, non-native accented speaker populations. With the state-of-the-art HuBERT model as a baseline, we propose and investigate self-supervised adaptation of speech representations to such populations in a parameter-efficient way via training accent-specific residual adapters. We experiment with 4 accents and choose automatic speech recognition (ASR) as the downstream task of interest. We obtain strong word error rate reductions (WERR) over HuBERT-large for all 4 accents, with a mean WERR of 22.7% with accent-specific adapters and a mean WERR of 25.1% if the entire encoder is accent-adapted. While our experiments utilize HuBERT and ASR as the downstream task, our proposed approach is both model and task-agnostic.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Anshu Bhatia (4 papers)
  2. Sanchit Sinha (11 papers)
  3. Saket Dingliwal (22 papers)
  4. Karthik Gopalakrishnan (34 papers)
  5. Sravan Bodapati (31 papers)
  6. Katrin Kirchhoff (36 papers)
Citations (4)