Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning From Yourself: A Self-Distillation Method for Fake Speech Detection (2303.01211v1)

Published 2 Mar 2023 in cs.SD, cs.LG, cs.MM, and eess.AS

Abstract: In this paper, we propose a novel self-distillation method for fake speech detection (FSD), which can significantly improve the performance of FSD without increasing the model complexity. For FSD, some fine-grained information is very important, such as spectrogram defects, mute segments, and so on, which are often perceived by shallow networks. However, shallow networks have much noise, which can not capture this very well. To address this problem, we propose using the deepest network instruct shallow network for enhancing shallow networks. Specifically, the networks of FSD are divided into several segments, the deepest network being used as the teacher model, and all shallow networks become multiple student models by adding classifiers. Meanwhile, the distillation path between the deepest network feature and shallow network features is used to reduce the feature difference. A series of experimental results on the ASVspoof 2019 LA and PA datasets show the effectiveness of the proposed method, with significant improvements compared to the baseline.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Jun Xue (19 papers)
  2. Cunhang Fan (35 papers)
  3. Jiangyan Yi (77 papers)
  4. Chenglong Wang (80 papers)
  5. Zhengqi Wen (69 papers)
  6. Dan Zhang (171 papers)
  7. Zhao Lv (22 papers)
Citations (31)

Summary

We haven't generated a summary for this paper yet.