Mutual-Learning Knowledge Distillation for Nighttime UAV Tracking (2312.07884v2)
Abstract: Nighttime unmanned aerial vehicle (UAV) tracking has been facilitated with indispensable plug-and-play low-light enhancers. However, the introduction of low-light enhancers increases the extra computational burden for the UAV, significantly hindering the development of real-time UAV applications. Meanwhile, these state-of-the-art (SOTA) enhancers lack tight coupling with the advanced daytime UAV tracking approach. To solve the above issues, this work proposes a novel mutual-learning knowledge distillation framework for nighttime UAV tracking, i.e., MLKD. This framework is constructed to learn a compact and fast nighttime tracker via knowledge transferring from the teacher and knowledge sharing among various students. Specifically, an advanced teacher based on a SOTA enhancer and a superior tracking backbone is adopted for guiding the student based only on the tight coupling-aware tracking backbone to directly extract nighttime object features. To address the biased learning of a single student, diverse lightweight students with different distillation methods are constructed to focus on various aspects of the teacher's knowledge. Moreover, an innovative mutual-learning room is designed to elect the superior student candidate to assist the remaining students frame-by-frame in the training phase. Furthermore, the final best student, i.e., MLKD-Track, is selected through the testing dataset. Extensive experiments demonstrate the effectiveness and superiority of MLKD and MLKD-Track. The practicality of the MLKD-Track is verified in real-world tests with different challenging situations. The code is available at https://github.com/lyfeng001/MLKD.
- C. Fu, K. Lu, G. Zheng, J. Ye, Z. Cao, B. Li, and G. Lu, “Siamese Object Tracking for Unmanned Aerial Vehicle: A Review and Comprehensive Analysis,” Artificial Intelligence Review, pp. 1–61, 2023.
- C. Ding, L. Lu, C. Wang, and C. Ding, “Design, Sensing, and Control of a Novel UAV Platform for Aerial Drilling and Screwing,” IEEE Robotics and Automation Letters, vol. 6, no. 2, pp. 3176–3183, 2021.
- Y. Li, C. Fu, F. Ding, Z. Huang, and G. Lu, “AutoTrack: Towards High-Performance Visual Tracking for UAV With Automatic Spatio-Temporal Regularization,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020, pp. 11 920–11 929.
- H. Fan, H. Bai, L. Lin, F. Yang, P. Chu, G. Deng, S. Yu, M. Huang, J. Liu, Y. Xu et al., “LaSOT: A High-Quality Large-Scale Single Object Tracking Benchmark,” International Journal of Computer Vision, vol. 129, pp. 439–461, 2021.
- L. Huang, X. Zhao, and K. Huang, “GOT-10k: A Large High-Diversity Benchmark for Generic Object Tracking in the Wild,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 43, no. 5, pp. 1562–1577, 2021.
- B. Li, W. Wu, Q. Wang, F. Zhang, J. Xing, and J. Yan, “SiamRPN++: Evolution of Siamese Visual Tracking with Very Deep Networks,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 4282–4291.
- Z. Cao, C. Fu, J. Ye, B. Li, and Y. Li, “HiFT: Hierarchical Feature Transformer for Aerial Tracking,” in IEEE International Conference on Computer Vision (ICCV), 2021, pp. 15 457–15 466.
- C. Fu, M. Cai, S. Li, K. Lu, H. Zuo, and C. Liu, “Continuity-Aware Latent Interframe Information Mining for Reliable UAV Tracking,” arXiv preprint arXiv:2303.04525, pp. 1–8, 2023.
- J. Ye, C. Fu, G. Zheng, Z. Cao, and B. Li, “DarkLighter: Light Up the Darkness for UAV Tracking,” in Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2021, pp. 3079–3085.
- J. Ye, C. Fu, Z. Cao, S. An, G. Zheng, and B. Li, “Tracker Meets Night: A Transformer Enhancer for UAV Tracking,” IEEE Robotics and Automation Letters, vol. 7, no. 2, pp. 3866–3873, 2022.
- C. Fu, H. Dong, J. Ye, G. Zheng, S. Li, and J. Zhao, “HighlightNet: Highlighting Low-Light Potential Features for Real-Time UAV Tracking,” in Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2022, pp. 12 146–12 153.
- R. Liu, L. Ma, J. Zhang, X. Fan, and Z. Luo, “Retinex-Inspired Unrolling with Cooperative Prior Architecture Search for Low-Light Image Enhancement,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021, pp. 10 561–10 570.
- Y. Jiang, X. Gong, D. Liu, Y. Cheng, C. Fang, X. Shen, J. Yang, P. Zhou, and Z. Wang, “EnlightenGAN: Deep Light Enhancement without Paired Supervision,” IEEE Transactions on Image Processing, vol. 30, pp. 2340–2349, 2021.
- F. Zhang, Y. Li, S. You, and Y. Fu, “Learning Temporal Consistency for Low Light Video Enhancement from Single Images,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021, pp. 4967–4976.
- C. Li, C. Guo, and C. C. Loy, “Learning to Enhance Low-Light Image via Zero-Reference Deep Curve Estimation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 44, no. 8, pp. 4225–4238, 2021.
- A. Romero, N. Ballas, S. E. Kahou, A. Chassang, C. Gatta, and Y. Bengio, “FitNets: Hints for Thin Deep Nets,” in Proceedings of the International Conference on Learning Representations (ICLR), 2015, pp. 1–13.
- Z. Zheng, R. Ye, P. Wang, D. Ren, W. Zuo, Q. Hou, and M.-M. Cheng, “Localization Distillation for Dense Object Detection,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, pp. 9407–9416.
- Y. Liu, K. Chen, C. Liu, Z. Qin, Z. Luo, and J. Wang, “Structured Knowledge Distillation for Semantic Segmentation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 2604–2613.
- G. Hinton, O. Vinyals, and J. Dean, “Distilling the Knowledge in a Neural Network,” arXiv preprint arXiv:1503.02531, pp. 1–9, 2015.
- H. Zuo, C. Fu, S. Li, J. Ye, and G. Zheng, “DeconNet: End-to-End Decontaminated Network for Vision-Based Aerial Tracking,” IEEE Transactions on Geoscience and Remote Sensing, vol. 60, pp. 1–12, 2022.
- G. Zheng, C. Fu, J. Ye, F. Lin, and F. Ding, “Mutation Sensitive Correlation Filter for Real-Time UAV Tracking with Adaptive Hybrid Label,” in Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), 2021, pp. 503–509.
- L. Bertinetto, J. Valmadre, J. F. Henriques, A. Vedaldi, and P. H. Torr, “Fully-Convolutional Siamese Networks for Object Tracking,” in Proceedings of the European Conference on Computer Vision (ECCV)Workshops, 2016, pp. 850–865.
- Z. Cao, C. Fu, J. Ye, B. Li, and Y. Li, “SiamAPN++: Siamese Attentional Aggregation Network for Real-Time UAV Tracking,” in Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, 2021, pp. 3086–3092.
- C. Fu, Z. Cao, Y. Li, J. Ye, and C. Feng, “Siamese Anchor Proposal Network for High-Speed Aerial Tracking,” in Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), 2021, pp. 510–516.
- B. Li, J. Yan, W. Wu, Z. Zhu, and X. Hu, “High Performance Visual Tracking with Siamese Region Proposal Network,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2018, pp. 8971–8980.
- H. Zuo, C. Fu, S. Li, J. Ye, and G. Zheng, “End-to-End Feature Decontaminated Network for UAV Tracking,” in Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2022, pp. 12 130–12 137.
- Z. Cao, Z. Huang, L. Pan, S. Zhang, Z. Liu, and C. Fu, “TCTrack: Temporal Contexts for Aerial Tracking,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, pp. 14 798–14 808.
- J. Ye, C. Fu, G. Zheng, D. P. Paudel, and G. Chen, “Unsupervised Domain Adaptation for Nighttime Aerial Tracking,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, pp. 8896–8905.
- L. Yao, H. Zuo, G. Zheng, C. Fu, and J. Pan, “SAM-DA: UAV Tracks Anything at Night with SAM-Powered Domain Adaptation,” arXiv preprint arXiv:2307.01024, pp. 1–12, 2023.
- C. Yang, H. Zhou, Z. An, X. Jiang, Y. Xu, and Q. Zhang, “Cross-Image Relational Knowledge Distillation for Semantic Segmentation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, pp. 12 319–12 328.
- A. Krizhevsky, I. Sutskever, and G. E. Hinton, “ImageNet Classification with Deep Convolutional Neural Networks,” Communications of the ACM, vol. 60, no. 6, pp. 84–90, 2017.
- B. Li, C. Fu, F. Ding, J. Ye, and F. Lin, “All-Day Object Tracking for Unmanned Aerial Vehicle,” IEEE Transactions on Mobile Computing, pp. 1–14, 2022.
- Y. Wu, J. Lim, and M.-H. Yang, “Online Object Tracking: A Benchmark,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2013, pp. 2411–2418.
- C. Guo, C. Li, J. Guo, C. C. Loy, J. Hou, S. Kwong, and R. Cong, “Zero-Reference Deep Curve Estimation for Low-Light Image Enhancement,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020, pp. 1780–1789.
- B. Yan, H. Peng, K. Wu, D. Wang, J. Fu, and H. Lu, “Lighttrack: Finding Lightweight Neural Networks for Object Tracking via One-Shot Architecture Search,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021, pp. 15 180–15 189.
- Y. Xu, Z. Wang, Z. Li, Y. Yuan, and G. Yu, “Siamfc++: Towards Robust and Accurate Visual Tracking with Target Estimation Guidelines,” in Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), vol. 34, no. 07, 2020, pp. 12 549–12 556.
- Z. Zhang and H. Peng, “Deeper and Wider Siamese Networks for Real-Time Visual Tracking,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 4591–4600.
- Z. Zhang, H. Peng, J. Fu, B. Li, and W. Hu, “Ocean: Object-Aware Anchor-Free Tracking,” in Proceedings of the European Conference on Computer Vision (ECCV). Springer, 2020, pp. 771–787.
- Z. Zhu, Q. Wang, B. Li, W. Wu, J. Yan, and W. Hu, “Distractor-Aware Siamese Networks for Visual Object Tracking,” in Proceedings of the European Conference on Computer Vision (ECCV), 2018, pp. 101–117.
- I. Sosnovik, A. Moskalev, and A. W. Smeulders, “Scale Equivariance Improves Siamese Tracking,” in Proceedings of the IEEE Winter Conference on Applications of Computer Vision (WACV), 2021, pp. 2765–2774.