Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

DPHuBERT: Joint Distillation and Pruning of Self-Supervised Speech Models (2305.17651v1)

Published 28 May 2023 in cs.CL, cs.SD, and eess.AS

Abstract: Self-supervised learning (SSL) has achieved notable success in many speech processing tasks, but the large model size and heavy computational cost hinder the deployment. Knowledge distillation trains a small student model to mimic the behavior of a large teacher model. However, the student architecture usually needs to be manually designed and will remain fixed during training, which requires prior knowledge and can lead to suboptimal performance. Inspired by recent success of task-specific structured pruning, we propose DPHuBERT, a novel task-agnostic compression method for speech SSL based on joint distillation and pruning. Experiments on SUPERB show that DPHuBERT outperforms pure distillation methods in almost all tasks. Moreover, DPHuBERT requires little training time and performs well with limited training data, making it suitable for resource-constrained applications. Our method can also be applied to various speech SSL models. Our code and models will be publicly available.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Yifan Peng (147 papers)
  2. Yui Sudo (15 papers)
  3. Shakeel Muhammad (2 papers)
  4. Shinji Watanabe (416 papers)
Citations (30)

Summary

We haven't generated a summary for this paper yet.