Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
169 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Mutual-Learning Knowledge Distillation for Nighttime UAV Tracking (2312.07884v2)

Published 13 Dec 2023 in cs.CV

Abstract: Nighttime unmanned aerial vehicle (UAV) tracking has been facilitated with indispensable plug-and-play low-light enhancers. However, the introduction of low-light enhancers increases the extra computational burden for the UAV, significantly hindering the development of real-time UAV applications. Meanwhile, these state-of-the-art (SOTA) enhancers lack tight coupling with the advanced daytime UAV tracking approach. To solve the above issues, this work proposes a novel mutual-learning knowledge distillation framework for nighttime UAV tracking, i.e., MLKD. This framework is constructed to learn a compact and fast nighttime tracker via knowledge transferring from the teacher and knowledge sharing among various students. Specifically, an advanced teacher based on a SOTA enhancer and a superior tracking backbone is adopted for guiding the student based only on the tight coupling-aware tracking backbone to directly extract nighttime object features. To address the biased learning of a single student, diverse lightweight students with different distillation methods are constructed to focus on various aspects of the teacher's knowledge. Moreover, an innovative mutual-learning room is designed to elect the superior student candidate to assist the remaining students frame-by-frame in the training phase. Furthermore, the final best student, i.e., MLKD-Track, is selected through the testing dataset. Extensive experiments demonstrate the effectiveness and superiority of MLKD and MLKD-Track. The practicality of the MLKD-Track is verified in real-world tests with different challenging situations. The code is available at https://github.com/lyfeng001/MLKD.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (40)
  1. C. Fu, K. Lu, G. Zheng, J. Ye, Z. Cao, B. Li, and G. Lu, “Siamese Object Tracking for Unmanned Aerial Vehicle: A Review and Comprehensive Analysis,” Artificial Intelligence Review, pp. 1–61, 2023.
  2. C. Ding, L. Lu, C. Wang, and C. Ding, “Design, Sensing, and Control of a Novel UAV Platform for Aerial Drilling and Screwing,” IEEE Robotics and Automation Letters, vol. 6, no. 2, pp. 3176–3183, 2021.
  3. Y. Li, C. Fu, F. Ding, Z. Huang, and G. Lu, “AutoTrack: Towards High-Performance Visual Tracking for UAV With Automatic Spatio-Temporal Regularization,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020, pp. 11 920–11 929.
  4. H. Fan, H. Bai, L. Lin, F. Yang, P. Chu, G. Deng, S. Yu, M. Huang, J. Liu, Y. Xu et al., “LaSOT: A High-Quality Large-Scale Single Object Tracking Benchmark,” International Journal of Computer Vision, vol. 129, pp. 439–461, 2021.
  5. L. Huang, X. Zhao, and K. Huang, “GOT-10k: A Large High-Diversity Benchmark for Generic Object Tracking in the Wild,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 43, no. 5, pp. 1562–1577, 2021.
  6. B. Li, W. Wu, Q. Wang, F. Zhang, J. Xing, and J. Yan, “SiamRPN++: Evolution of Siamese Visual Tracking with Very Deep Networks,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 4282–4291.
  7. Z. Cao, C. Fu, J. Ye, B. Li, and Y. Li, “HiFT: Hierarchical Feature Transformer for Aerial Tracking,” in IEEE International Conference on Computer Vision (ICCV), 2021, pp. 15 457–15 466.
  8. C. Fu, M. Cai, S. Li, K. Lu, H. Zuo, and C. Liu, “Continuity-Aware Latent Interframe Information Mining for Reliable UAV Tracking,” arXiv preprint arXiv:2303.04525, pp. 1–8, 2023.
  9. J. Ye, C. Fu, G. Zheng, Z. Cao, and B. Li, “DarkLighter: Light Up the Darkness for UAV Tracking,” in Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2021, pp. 3079–3085.
  10. J. Ye, C. Fu, Z. Cao, S. An, G. Zheng, and B. Li, “Tracker Meets Night: A Transformer Enhancer for UAV Tracking,” IEEE Robotics and Automation Letters, vol. 7, no. 2, pp. 3866–3873, 2022.
  11. C. Fu, H. Dong, J. Ye, G. Zheng, S. Li, and J. Zhao, “HighlightNet: Highlighting Low-Light Potential Features for Real-Time UAV Tracking,” in Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2022, pp. 12 146–12 153.
  12. R. Liu, L. Ma, J. Zhang, X. Fan, and Z. Luo, “Retinex-Inspired Unrolling with Cooperative Prior Architecture Search for Low-Light Image Enhancement,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021, pp. 10 561–10 570.
  13. Y. Jiang, X. Gong, D. Liu, Y. Cheng, C. Fang, X. Shen, J. Yang, P. Zhou, and Z. Wang, “EnlightenGAN: Deep Light Enhancement without Paired Supervision,” IEEE Transactions on Image Processing, vol. 30, pp. 2340–2349, 2021.
  14. F. Zhang, Y. Li, S. You, and Y. Fu, “Learning Temporal Consistency for Low Light Video Enhancement from Single Images,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021, pp. 4967–4976.
  15. C. Li, C. Guo, and C. C. Loy, “Learning to Enhance Low-Light Image via Zero-Reference Deep Curve Estimation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 44, no. 8, pp. 4225–4238, 2021.
  16. A. Romero, N. Ballas, S. E. Kahou, A. Chassang, C. Gatta, and Y. Bengio, “FitNets: Hints for Thin Deep Nets,” in Proceedings of the International Conference on Learning Representations (ICLR), 2015, pp. 1–13.
  17. Z. Zheng, R. Ye, P. Wang, D. Ren, W. Zuo, Q. Hou, and M.-M. Cheng, “Localization Distillation for Dense Object Detection,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, pp. 9407–9416.
  18. Y. Liu, K. Chen, C. Liu, Z. Qin, Z. Luo, and J. Wang, “Structured Knowledge Distillation for Semantic Segmentation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 2604–2613.
  19. G. Hinton, O. Vinyals, and J. Dean, “Distilling the Knowledge in a Neural Network,” arXiv preprint arXiv:1503.02531, pp. 1–9, 2015.
  20. H. Zuo, C. Fu, S. Li, J. Ye, and G. Zheng, “DeconNet: End-to-End Decontaminated Network for Vision-Based Aerial Tracking,” IEEE Transactions on Geoscience and Remote Sensing, vol. 60, pp. 1–12, 2022.
  21. G. Zheng, C. Fu, J. Ye, F. Lin, and F. Ding, “Mutation Sensitive Correlation Filter for Real-Time UAV Tracking with Adaptive Hybrid Label,” in Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), 2021, pp. 503–509.
  22. L. Bertinetto, J. Valmadre, J. F. Henriques, A. Vedaldi, and P. H. Torr, “Fully-Convolutional Siamese Networks for Object Tracking,” in Proceedings of the European Conference on Computer Vision (ECCV)Workshops, 2016, pp. 850–865.
  23. Z. Cao, C. Fu, J. Ye, B. Li, and Y. Li, “SiamAPN++: Siamese Attentional Aggregation Network for Real-Time UAV Tracking,” in Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).   IEEE, 2021, pp. 3086–3092.
  24. C. Fu, Z. Cao, Y. Li, J. Ye, and C. Feng, “Siamese Anchor Proposal Network for High-Speed Aerial Tracking,” in Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), 2021, pp. 510–516.
  25. B. Li, J. Yan, W. Wu, Z. Zhu, and X. Hu, “High Performance Visual Tracking with Siamese Region Proposal Network,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2018, pp. 8971–8980.
  26. H. Zuo, C. Fu, S. Li, J. Ye, and G. Zheng, “End-to-End Feature Decontaminated Network for UAV Tracking,” in Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2022, pp. 12 130–12 137.
  27. Z. Cao, Z. Huang, L. Pan, S. Zhang, Z. Liu, and C. Fu, “TCTrack: Temporal Contexts for Aerial Tracking,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, pp. 14 798–14 808.
  28. J. Ye, C. Fu, G. Zheng, D. P. Paudel, and G. Chen, “Unsupervised Domain Adaptation for Nighttime Aerial Tracking,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, pp. 8896–8905.
  29. L. Yao, H. Zuo, G. Zheng, C. Fu, and J. Pan, “SAM-DA: UAV Tracks Anything at Night with SAM-Powered Domain Adaptation,” arXiv preprint arXiv:2307.01024, pp. 1–12, 2023.
  30. C. Yang, H. Zhou, Z. An, X. Jiang, Y. Xu, and Q. Zhang, “Cross-Image Relational Knowledge Distillation for Semantic Segmentation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, pp. 12 319–12 328.
  31. A. Krizhevsky, I. Sutskever, and G. E. Hinton, “ImageNet Classification with Deep Convolutional Neural Networks,” Communications of the ACM, vol. 60, no. 6, pp. 84–90, 2017.
  32. B. Li, C. Fu, F. Ding, J. Ye, and F. Lin, “All-Day Object Tracking for Unmanned Aerial Vehicle,” IEEE Transactions on Mobile Computing, pp. 1–14, 2022.
  33. Y. Wu, J. Lim, and M.-H. Yang, “Online Object Tracking: A Benchmark,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2013, pp. 2411–2418.
  34. C. Guo, C. Li, J. Guo, C. C. Loy, J. Hou, S. Kwong, and R. Cong, “Zero-Reference Deep Curve Estimation for Low-Light Image Enhancement,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020, pp. 1780–1789.
  35. B. Yan, H. Peng, K. Wu, D. Wang, J. Fu, and H. Lu, “Lighttrack: Finding Lightweight Neural Networks for Object Tracking via One-Shot Architecture Search,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021, pp. 15 180–15 189.
  36. Y. Xu, Z. Wang, Z. Li, Y. Yuan, and G. Yu, “Siamfc++: Towards Robust and Accurate Visual Tracking with Target Estimation Guidelines,” in Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), vol. 34, no. 07, 2020, pp. 12 549–12 556.
  37. Z. Zhang and H. Peng, “Deeper and Wider Siamese Networks for Real-Time Visual Tracking,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 4591–4600.
  38. Z. Zhang, H. Peng, J. Fu, B. Li, and W. Hu, “Ocean: Object-Aware Anchor-Free Tracking,” in Proceedings of the European Conference on Computer Vision (ECCV).   Springer, 2020, pp. 771–787.
  39. Z. Zhu, Q. Wang, B. Li, W. Wu, J. Yan, and W. Hu, “Distractor-Aware Siamese Networks for Visual Object Tracking,” in Proceedings of the European Conference on Computer Vision (ECCV), 2018, pp. 101–117.
  40. I. Sosnovik, A. Moskalev, and A. W. Smeulders, “Scale Equivariance Improves Siamese Tracking,” in Proceedings of the IEEE Winter Conference on Applications of Computer Vision (WACV), 2021, pp. 2765–2774.

Summary

We haven't generated a summary for this paper yet.