Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Label-free Knowledge Distillation with Contrastive Loss for Light-weight Speaker Recognition (2212.03090v1)

Published 6 Dec 2022 in cs.SD and eess.AS

Abstract: Very deep models for speaker recognition (SR) have demonstrated remarkable performance improvement in recent research. However, it is impractical to deploy these models for on-device applications with constrained computational resources. On the other hand, light-weight models are highly desired in practice despite their sub-optimal performance. This research aims to improve light-weight SR models through large-scale label-free knowledge distillation (KD). Existing KD approaches for SR typically require speaker labels to learn task-specific knowledge, due to the inefficiency of conventional loss for distillation. To address the inefficiency problem and achieve label-free KD, we propose to employ the contrastive loss from self-supervised learning for distillation. Extensive experiments are conducted on a collection of public speech datasets from diverse sources. Results on light-weight SR models show that the proposed approach of label-free KD with contrastive loss consistently outperforms both conventional distillation methods and self-supervised learning methods by a significant margin.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Zhiyuan Peng (33 papers)
  2. Xuanji He (4 papers)
  3. Ke Ding (30 papers)
  4. Tan Lee (70 papers)
  5. Guanglu Wan (24 papers)
Citations (4)

Summary

We haven't generated a summary for this paper yet.