Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Dynamic Data Pruning for Automatic Speech Recognition (2406.18373v1)

Published 26 Jun 2024 in cs.CL, cs.SD, and eess.AS

Abstract: The recent success of Automatic Speech Recognition (ASR) is largely attributed to the ever-growing amount of training data. However, this trend has made model training prohibitively costly and imposed computational demands. While data pruning has been proposed to mitigate this issue by identifying a small subset of relevant data, its application in ASR has been barely explored, and existing works often entail significant overhead to achieve meaningful results. To fill this gap, this paper presents the first investigation of dynamic data pruning for ASR, finding that we can reach the full-data performance by dynamically selecting 70% of data. Furthermore, we introduce Dynamic Data Pruning for ASR (DDP-ASR), which offers several fine-grained pruning granularities specifically tailored for speech-related datasets, going beyond the conventional pruning of entire time sequences. Our intensive experiments show that DDP-ASR can save up to 1.6x training time with negligible performance loss.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Qiao Xiao (14 papers)
  2. Pingchuan Ma (91 papers)
  3. Adriana Fernandez-Lopez (7 papers)
  4. Boqian Wu (9 papers)
  5. Lu Yin (85 papers)
  6. Stavros Petridis (64 papers)
  7. Mykola Pechenizkiy (118 papers)
  8. Maja Pantic (100 papers)
  9. Decebal Constantin Mocanu (52 papers)
  10. Shiwei Liu (76 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.