Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Spirit Distillation: A Model Compression Method with Multi-domain Knowledge Transfer (2104.14696v1)

Published 29 Apr 2021 in cs.CV

Abstract: Recent applications pose requirements of both cross-domain knowledge transfer and model compression to machine learning models due to insufficient training data and limited computational resources. In this paper, we propose a new knowledge distillation model, named Spirit Distillation (SD), which is a model compression method with multi-domain knowledge transfer. The compact student network mimics out a representation equivalent to the front part of the teacher network, through which the general knowledge can be transferred from the source domain (teacher) to the target domain (student). To further improve the robustness of the student, we extend SD to Enhanced Spirit Distillation (ESD) in exploiting a more comprehensive knowledge by introducing the proximity domain which is similar to the target domain for feature extraction. Results demonstrate that our method can boost mIOU and high-precision accuracy by 1.4% and 8.2% respectively with 78.2% segmentation variance, and can gain a precise compact network with only 41.8% FLOPs.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Zhiyuan Wu (34 papers)
  2. Yu Jiang (166 papers)
  3. Minghao Zhao (17 papers)
  4. Chupeng Cui (3 papers)
  5. Zongmin Yang (3 papers)
  6. Xinhui Xue (3 papers)
  7. Hong Qi (25 papers)
Citations (9)

Summary

We haven't generated a summary for this paper yet.