Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Hyper-parameter Adaptation of Conformer ASR Systems for Elderly and Dysarthric Speech Recognition (2306.15265v1)

Published 27 Jun 2023 in eess.AS and cs.LG

Abstract: Automatic recognition of disordered and elderly speech remains highly challenging tasks to date due to data scarcity. Parameter fine-tuning is often used to exploit the large quantities of non-aged and healthy speech pre-trained models, while neural architecture hyper-parameters are set using expert knowledge and remain unchanged. This paper investigates hyper-parameter adaptation for Conformer ASR systems that are pre-trained on the Librispeech corpus before being domain adapted to the DementiaBank elderly and UASpeech dysarthric speech datasets. Experimental results suggest that hyper-parameter adaptation produced word error rate (WER) reductions of 0.45% and 0.67% over parameter-only fine-tuning on DBank and UASpeech tasks respectively. An intuitive correlation is found between the performance improvements by hyper-parameter domain adaptation and the relative utterance length ratio between the source and target domain data.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Tianzi Wang (37 papers)
  2. Shoukang Hu (38 papers)
  3. Jiajun Deng (75 papers)
  4. Zengrui Jin (30 papers)
  5. Mengzhe Geng (42 papers)
  6. Yi Wang (1038 papers)
  7. Helen Meng (204 papers)
  8. Xunying Liu (92 papers)
Citations (4)

Summary

We haven't generated a summary for this paper yet.