Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

LumiNet: The Bright Side of Perceptual Knowledge Distillation (2310.03669v2)

Published 5 Oct 2023 in cs.CV

Abstract: In knowledge distillation literature, feature-based methods have dominated due to their ability to effectively tap into extensive teacher models. In contrast, logit-based approaches, which aim to distill `dark knowledge' from teachers, typically exhibit inferior performance compared to feature-based methods. To bridge this gap, we present LumiNet, a novel knowledge distillation algorithm designed to enhance logit-based distillation. We introduce the concept of 'perception', aiming to calibrate logits based on the model's representation capability. This concept addresses overconfidence issues in logit-based distillation method while also introducing a novel method to distill knowledge from the teacher. It reconstructs the logits of a sample/instances by considering relationships with other samples in the batch. LumiNet excels on benchmarks like CIFAR-100, ImageNet, and MSCOCO, outperforming leading feature-based methods, e.g., compared to KD with ResNet18 and MobileNetV2 on ImageNet, it shows improvements of 1.5% and 2.05%, respectively.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Md. Ismail Hossain (5 papers)
  2. M M Lutfe Elahi (4 papers)
  3. Sameera Ramasinghe (36 papers)
  4. Ali Cheraghian (18 papers)
  5. Fuad Rahman (12 papers)
  6. Nabeel Mohammed (27 papers)
  7. Shafin Rahman (38 papers)