Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

LumiNet: The Bright Side of Perceptual Knowledge Distillation (2310.03669v2)

Published 5 Oct 2023 in cs.CV

Abstract: In knowledge distillation literature, feature-based methods have dominated due to their ability to effectively tap into extensive teacher models. In contrast, logit-based approaches, which aim to distill `dark knowledge' from teachers, typically exhibit inferior performance compared to feature-based methods. To bridge this gap, we present LumiNet, a novel knowledge distillation algorithm designed to enhance logit-based distillation. We introduce the concept of 'perception', aiming to calibrate logits based on the model's representation capability. This concept addresses overconfidence issues in logit-based distillation method while also introducing a novel method to distill knowledge from the teacher. It reconstructs the logits of a sample/instances by considering relationships with other samples in the batch. LumiNet excels on benchmarks like CIFAR-100, ImageNet, and MSCOCO, outperforming leading feature-based methods, e.g., compared to KD with ResNet18 and MobileNetV2 on ImageNet, it shows improvements of 1.5% and 2.05%, respectively.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Md. Ismail Hossain (5 papers)
  2. M M Lutfe Elahi (4 papers)
  3. Sameera Ramasinghe (36 papers)
  4. Ali Cheraghian (18 papers)
  5. Fuad Rahman (12 papers)
  6. Nabeel Mohammed (27 papers)
  7. Shafin Rahman (38 papers)

Summary

We haven't generated a summary for this paper yet.