Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

I2CKD : Intra- and Inter-Class Knowledge Distillation for Semantic Segmentation (2403.18490v1)

Published 27 Mar 2024 in cs.CV

Abstract: This paper proposes a new knowledge distillation method tailored for image semantic segmentation, termed Intra- and Inter-Class Knowledge Distillation (I2CKD). The focus of this method is on capturing and transferring knowledge between the intermediate layers of teacher (cumbersome model) and student (compact model). For knowledge extraction, we exploit class prototypes derived from feature maps. To facilitate knowledge transfer, we employ a triplet loss in order to minimize intra-class variances and maximize inter-class variances between teacher and student prototypes. Consequently, I2CKD enables the student to better mimic the feature representation of the teacher for each class, thereby enhancing the segmentation performance of the compact network. Extensive experiments on three segmentation datasets, i.e., Cityscapes, Pascal VOC and CamVid, using various teacher-student network pairs demonstrate the effectiveness of the proposed method.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (29)
  1. “Survey on semantic segmentation using deep learning techniques,” Neurocomputing, vol. 338, pp. 321–348, 2019.
  2. “Espnet: Efficient spatial pyramid of dilated convolutions for semantic segmentation,” in Proceedings of the european conference on computer vision (ECCV), 2018, pp. 552–568.
  3. “Enet: A deep neural network architecture for real-time semantic segmentation,” arXiv preprint arXiv:1606.02147, 2016.
  4. “Bisenet: Bilateral segmentation network for real-time semantic segmentation,” in Proceedings of the European conference on computer vision (ECCV), 2018, pp. 325–341.
  5. “What is the state of neural network pruning?,” Proceedings of machine learning and systems, vol. 2, pp. 129–146, 2020.
  6. “Quantized convolutional neural networks for mobile devices,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 4820–4828.
  7. “Distilling the knowledge in a neural network,” arXiv preprint arXiv:1503.02531, vol. 2, no. 7, 2015.
  8. “Categories of response-based, feature-based, and relation-based knowledge distillation,” in Advancements in Knowledge Distillation: Towards New Horizons of Intelligent Systems, pp. 1–32. Springer, 2023.
  9. “Improving fast segmentation with teacher-student learning,” in British Machine Vision Conference 2018, BMVC 2018, Newcastle, UK, September 3-6, 2018. 2018, p. 205, BMVA Press.
  10. “Structured knowledge distillation for dense prediction,” IEEE transactions on pattern analysis and machine intelligence, 2020.
  11. “Intra-class feature variation distillation for semantic segmentation,” in European Conference on Computer Vision. Springer, 2020, pp. 346–362.
  12. “Knowledge adaptation for efficient semantic segmentation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 578–587.
  13. “Knowledge distillation for semantic segmentation using channel and spatial correlations and adaptive cross entropy,” Sensors, vol. 20, no. 16, pp. 4616, 2020.
  14. “Channel-wise knowledge distillation for dense prediction,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 5311–5320.
  15. “Exploring inter-channel correlation for diversity-preserved knowledge distillation,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 8271–8280.
  16. “Cross-image relational knowledge distillation for semantic segmentation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 12319–12328.
  17. “Knowledge distillation: A survey,” International Journal of Computer Vision, vol. 129, no. 6, pp. 1789–1819, 2021.
  18. “Fitnets: Hints for thin deep nets,” in 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings, Yoshua Bengio and Yann LeCun, Eds., 2015.
  19. “Paying more attention to attention: Improving the performance of convolutional neural networks via attention transfer,” in 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings. 2017, OpenReview.net.
  20. “A gift from knowledge distillation: Fast optimization, network minimization and transfer learning,” in 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, July 21-26, 2017. 2017, pp. 7130–7138, IEEE Computer Society.
  21. “Relational knowledge distillation,” in IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, June 16-20, 2019. 2019, pp. 3967–3976, Computer Vision Foundation / IEEE.
  22. “Knowledge distillation via instance relationship graph,” in IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, June 16-20, 2019. 2019, pp. 7096–7104, Computer Vision Foundation / IEEE.
  23. “Fully convolutional networks for semantic segmentation,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 3431–3440.
  24. “Pyramid scene parsing network,” in 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017, pp. 6230–6239.
  25. “Rethinking atrous convolution for semantic image segmentation,” arXiv preprint arXiv:1706.05587, 2017.
  26. “Semantic images segmentation for autonomous driving using self-attention knowledge distillation,” in 16th International Conference on Signal-Image Technology & Internet-Based Systems, SITIS 2022, Dijon, France, October 19-21, 2022, Kokou Yétongnon, Albert Dipanda, and Luigi Gallo, Eds. 2022, pp. 198–202, IEEE.
  27. “The cityscapes dataset for semantic urban scene understanding,” in Proc. of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016.
  28. “The pascal visual object classes (voc) challenge,” International journal of computer vision, vol. 88, pp. 303–338, 2010.
  29. “Segmentation and recognition using structure from motion point clouds,” in Computer Vision – ECCV 2008, David Forsyth, Philip Torr, and Andrew Zisserman, Eds., Berlin, Heidelberg, 2008, pp. 44–57, Springer Berlin Heidelberg.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Ayoub Karine (2 papers)
  2. Maher Jridi (1 paper)
  3. Thibault Napoléon (1 paper)

Summary

We haven't generated a summary for this paper yet.