Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

MixSKD: Self-Knowledge Distillation from Mixup for Image Recognition (2208.05768v1)

Published 11 Aug 2022 in cs.CV

Abstract: Unlike the conventional Knowledge Distillation (KD), Self-KD allows a network to learn knowledge from itself without any guidance from extra networks. This paper proposes to perform Self-KD from image Mixture (MixSKD), which integrates these two techniques into a unified framework. MixSKD mutually distills feature maps and probability distributions between the random pair of original images and their mixup images in a meaningful way. Therefore, it guides the network to learn cross-image knowledge by modelling supervisory signals from mixup images. Moreover, we construct a self-teacher network by aggregating multi-stage feature maps for providing soft labels to supervise the backbone classifier, further improving the efficacy of self-boosting. Experiments on image classification and transfer learning to object detection and semantic segmentation demonstrate that MixSKD outperforms other state-of-the-art Self-KD and data augmentation methods. The code is available at https://github.com/winycg/Self-KD-Lib.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Chuanguang Yang (36 papers)
  2. Zhulin An (43 papers)
  3. Helong Zhou (9 papers)
  4. Linhang Cai (6 papers)
  5. Xiang Zhi (2 papers)
  6. Jiwen Wu (1 paper)
  7. Yongjun Xu (81 papers)
  8. Qian Zhang (308 papers)
Citations (40)