Low-Light Object Tracking: A Benchmark (2408.11463v1)
Abstract: In recent years, the field of visual tracking has made significant progress with the application of large-scale training datasets. These datasets have supported the development of sophisticated algorithms, enhancing the accuracy and stability of visual object tracking. However, most research has primarily focused on favorable illumination circumstances, neglecting the challenges of tracking in low-ligh environments. In low-light scenes, lighting may change dramatically, targets may lack distinct texture features, and in some scenarios, targets may not be directly observable. These factors can lead to a severe decline in tracking performance. To address this issue, we introduce LLOT, a benchmark specifically designed for Low-Light Object Tracking. LLOT comprises 269 challenging sequences with a total of over 132K frames, each carefully annotated with bounding boxes. This specially designed dataset aims to promote innovation and advancement in object tracking techniques for low-light conditions, addressing challenges not adequately covered by existing benchmarks. To assess the performance of existing methods on LLOT, we conducted extensive tests on 39 state-of-the-art tracking algorithms. The results highlight a considerable gap in low-light tracking performance. In response, we propose H-DCPT, a novel tracker that incorporates historical and darkness clue prompts to set a stronger baseline. H-DCPT outperformed all 39 evaluated methods in our experiments, demonstrating significant improvements. We hope that our benchmark and H-DCPT will stimulate the development of novel and accurate methods for tracking objects in low-light conditions. The LLOT and code are available at https://github.com/OpenCodeGithub/H-DCPT.
- J. R. SM and G. Augasta, “Review of recent advances in visual tracking techniques,” Multimedia tools and applications, vol. 80, pp. 24 185–24 203, 2021.
- J. Zhang, Z. Liu, and Y. Lin, “Correlation gaussian particle filter for robust visual tracking,” in 2020 Chinese Control And Decision Conference (CCDC). IEEE, 2020, pp. 4854–4857.
- W. Lijia, W. Binbin, and C. Xufeng, “An adaptive kernel based correlation filter algorithm for real time object tracking,” in 2020 12th International Conference on Intelligent Human-Machine Systems and Cybernetics (IHMSC), vol. 2. IEEE, 2020, pp. 144–148.
- M. Danelljan, L. V. Gool, and R. Timofte, “Probabilistic regression for visual tracking,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 7183–7192.
- S. M. Marvasti-Zadeh, L. Cheng, H. Ghanei-Yakhdan, and S. Kasaei, “Deep learning for visual tracking: A comprehensive survey,” IEEE Transactions on Intelligent Transportation Systems, vol. 23, no. 5, pp. 3943–3968, 2021.
- S. Gao, C. Zhou, C. Ma, X. Wang, and J. Yuan, “Aiatrack: Attention in attention for transformer visual tracking,” in European Conference on Computer Vision. Springer, 2022, pp. 146–164.
- D. Yuan, X. Chang, Q. Liu, Y. Yang, D. Wang, M. Shu, Z. He, and G. Shi, “Active learning for deep visual tracking,” IEEE Transactions on Neural Networks and Learning Systems, 2023.
- H. Zhu, M. Xue, Y. Wang, G. Yuan, and X. Li, “Fast visual tracking with siamese oriented region proposal network,” IEEE signal processing letters, vol. 29, pp. 1437–1441, 2022.
- T. Liu and Y. Liu, “Moving camera-based object tracking using adaptive ground plane estimation and constrained multiple kernels,” Journal of Advanced Transportation, vol. 2021, no. 1, p. 8153474, 2021.
- D. Liu, Y. Cui, W. Tan, and Y. Chen, “Sg-net: Spatial granularity network for one-stage video instance segmentation,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2021, pp. 9816–9825.
- F. Tang and Q. Ling, “Ranking-based siamese visual tracking,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2022, pp. 8741–8750.
- Z. Zhang, F. Wu, Y. Qiu, J. Liang, and S. Li, “Tracking small and fast moving objects: A benchmark,” in Proceedings of the Asian Conference on Computer Vision, 2022, pp. 4514–4530.
- Y. Wu, Y. Wang, Y. Liao, F. Wu, H. Ye, and S. Li, “Tracking transforming objects: A benchmark,” arXiv preprint arXiv:2404.18143, 2024.
- J. Zhu, H. Tang, Z.-Q. Cheng, J.-Y. He, B. Luo, S. Qiu, S. Li, and H. Lu, “Dcpt: Darkness clue-prompted tracking in nighttime uavs,” arXiv preprint arXiv:2309.10491, 2023.
- B. Li, C. Fu, F. Ding, J. Ye, and F. Lin, “All-day object tracking for unmanned aerial vehicle,” IEEE Transactions on Mobile Computing, vol. 22, no. 8, pp. 4515–4529, 2022.
- G. Zhang, Y. Zhang, X. Yuan, and Y. Fu, “Binarized low-light raw video enhancement,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2024, pp. 25 753–25 762.
- L. Sun, S. Kong, Z. Yang, D. Gao, and B. Fan, “Modified siamese network based on feature enhancement and dynamic template for low-light object tracking in uav videos,” Drones, vol. 7, no. 7, p. 483, 2023.
- Z. Fu, Y. Yang, X. Tu, Y. Huang, X. Ding, and K.-K. Ma, “Learning a simple low-light image enhancer from paired low-light instances,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2023, pp. 22 252–22 261.
- K. Robert, “Night-time traffic surveillance: A robust framework for multi-vehicle detection, classification and tracking,” in 2009 Sixth IEEE International Conference on Advanced Video and Signal Based Surveillance. IEEE, 2009, pp. 1–6.
- ——, “Video-based traffic monitoring at day and night vehicle features detection tracking,” in 2009 12th International IEEE Conference on Intelligent Transportation Systems. IEEE, 2009, pp. 1–6.
- Y.-L. Chen, B.-F. Wu, H.-Y. Huang, and C.-J. Fan, “A real-time vision system for nighttime vehicle detection and traffic surveillance,” IEEE Transactions on Industrial Electronics, vol. 58, no. 5, pp. 2030–2044, 2010.
- D. Chatziparaschis, M. G. Lagoudakis, and P. Partsinevelos, “Aerial and ground robot collaboration for autonomous mapping in search and rescue missions,” Drones, vol. 4, no. 4, p. 79, 2020.
- P. Tang, J. Li, and H. Sun, “A review of electric uav visual detection and navigation technologies for emergency rescue missions,” Sustainability, vol. 16, no. 5, p. 2105, 2024.
- J. P. Queralta, J. Taipalmaa, B. C. Pullinen, V. K. Sarker, T. N. Gia, H. Tenhunen, M. Gabbouj, J. Raitoharju, and T. Westerlund, “Collaborative multi-robot search and rescue: Planning, coordination, perception, and active vision,” Ieee Access, vol. 8, pp. 191 617–191 643, 2020.
- L. F. Hughey, A. M. Hein, A. Strandburg-Peshkin, and F. H. Jensen, “Challenges and solutions for studying collective animal behaviour in the wild,” Philosophical Transactions of the Royal Society B: Biological Sciences, vol. 373, no. 1746, p. 20170005, 2018.
- D. Tuia, B. Kellenberger, S. Beery, B. R. Costelloe, S. Zuffi, B. Risse, A. Mathis, M. W. Mathis, F. Van Langevelde, T. Burghardt et al., “Perspectives in machine learning for wildlife conservation,” Nature communications, vol. 13, no. 1, pp. 1–15, 2022.
- S. Abba, A. M. Bizi, J.-A. Lee, S. Bakouri, and M. L. Crespo, “Real-time object detection, tracking, and monitoring framework for security surveillance systems,” Heliyon, 2024.
- I. Ahmed and G. Jeon, “A real-time person tracking system based on siammask network for intelligent video surveillance,” Journal of Real-Time Image Processing, vol. 18, no. 5, pp. 1803–1814, 2021.
- N. Dilshad, J. Hwang, J. Song, and N. Sung, “Applications and challenges in video surveillance via drone: A brief survey,” in 2020 International Conference on Information and Communication Technology Convergence (ICTC). IEEE, 2020, pp. 728–732.
- W. Budiharto, E. Irwansyah, J. S. Suroso, and A. A. S. Gunawan, “Design of object tracking for military robot using pid controller and computer vision,” ICIC Express Letters, vol. 14, no. 3, pp. 289–294, 2020.
- T. Celik and T. Tjahjadi, “Contextual and variational contrast enhancement,” IEEE Transactions on Image Processing, vol. 20, no. 12, pp. 3431–3441, 2011.
- C. Lee, C. Lee, and C.-S. Kim, “Contrast enhancement based on layered difference representation of 2d histograms,” IEEE transactions on image processing, vol. 22, no. 12, pp. 5372–5384, 2013.
- X. Guo, Y. Li, and H. Ling, “Lime: Low-light image enhancement via illumination map estimation,” IEEE Transactions on image processing, vol. 26, no. 2, pp. 982–993, 2016.
- E. H. Land, “The retinex theory of color vision,” Scientific american, vol. 237, no. 6, pp. 108–129, 1977.
- C. Wei, W. Wang, W. Yang, and J. Liu, “Deep retinex decomposition for low-light enhancement,” arXiv preprint arXiv:1808.04560, 2018.
- R. Wang, Q. Zhang, C.-W. Fu, X. Shen, W.-S. Zheng, and J. Jia, “Underexposed photo enhancement using deep illumination estimation,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 6849–6857.
- Y. Jiang, X. Gong, D. Liu, Y. Cheng, C. Fang, X. Shen, J. Yang, P. Zhou, and Z. Wang, “Enlightengan: Deep light enhancement without paired supervision,” IEEE transactions on image processing, vol. 30, pp. 2340–2349, 2021.
- C. Fu, B. Li, F. Ding, F. Lin, and G. Lu, “Correlation filters for unmanned aerial vehicle-based aerial tracking: A review and experimental evaluation,” IEEE Geoscience and Remote Sensing Magazine, vol. 10, no. 1, pp. 125–160, 2021.
- S. Li, Y. Liu, Q. Zhao, and Z. Feng, “Learning residue-aware correlation filters and refining scale for real-time uav tracking,” Pattern Recognition, vol. 127, p. 108614, 2022.
- Y. Li, C. Fu, F. Ding, Z. Huang, and G. Lu, “Autotrack: Towards high-performance visual tracking for uav with automatic spatio-temporal regularization,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 11 923–11 932.
- Z. Huang, C. Fu, Y. Li, F. Lin, and P. Lu, “Learning aberrance repressed correlation filters for real-time uav tracking,” in Proceedings of the IEEE/CVF international conference on computer vision, 2019, pp. 2891–2900.
- Z. Cao, C. Fu, J. Ye, B. Li, and Y. Li, “Hift: Hierarchical feature transformer for aerial tracking,” in Proceedings of the IEEE/CVF international conference on computer vision, 2021, pp. 15 457–15 466.
- S. Li, Y. Yang, D. Zeng, and X. Wang, “Adaptive and background-aware vision transformer for real-time uav tracking,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023, pp. 13 989–14 000.
- J.-P. Lan, Z.-Q. Cheng, J.-Y. He, C. Li, B. Luo, X. Bao, W. Xiang, Y. Geng, and X. Xie, “Procontext: Exploring progressive context transformer for tracking,” in ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2023, pp. 1–5.
- W. Cai, Q. Liu, and Y. Wang, “Hiptrack: Visual tracking with historical prompts,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2024, pp. 19 258–19 267.
- B. Li, W. Wu, Q. Wang, F. Zhang, J. Xing, and J. Yan, “Siamrpn++: Evolution of siamese visual tracking with very deep networks,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 4282–4291.
- D. Guo, J. Wang, Y. Cui, Z. Wang, and S. Chen, “Siamcar: Siamese fully convolutional classification and regression for visual tracking,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 6269–6277.
- Y. Xu, Z. Wang, Z. Li, Y. Yuan, and G. Yu, “Siamfc++: Towards robust and accurate visual tracking with target estimation guidelines,” in Proceedings of the AAAI conference on artificial intelligence, vol. 34, no. 07, 2020, pp. 12 549–12 556.
- Z. Cao, C. Fu, J. Ye, B. Li, and Y. Li, “Siamapn++: Siamese attentional aggregation network for real-time uav tracking,” in 2021 IEEE/RSJ international conference on intelligent robots and systems (IROS). IEEE, 2021, pp. 3086–3092.
- D. Guo, Y. Shao, Y. Cui, Z. Wang, L. Zhang, and C. Shen, “Graph attention tracking,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2021, pp. 9543–9552.
- Z. Cao, Z. Huang, L. Pan, S. Zhang, Z. Liu, and C. Fu, “Tctrack: Temporal contexts for aerial tracking,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2022, pp. 14 798–14 808.
- B. Ye, H. Chang, B. Ma, S. Shan, and X. Chen, “Joint feature learning and relation modeling for tracking: A one-stream framework,” in European Conference on Computer Vision. Springer, 2022, pp. 341–357.
- J. Ye, C. Fu, G. Zheng, D. P. Paudel, and G. Chen, “Unsupervised domain adaptation for nighttime aerial tracking,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2022, pp. 8896–8905.
- J. Ye, C. Fu, G. Zheng, Z. Cao, and B. Li, “Darklighter: Light up the darkness for uav tracking,” in 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, 2021, pp. 3079–3085.
- C. Fu, H. Dong, J. Ye, G. Zheng, S. Li, and J. Zhao, “Highlightnet: highlighting low-light potential features for real-time uav tracking,” in 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, 2022, pp. 12 146–12 153.
- J. Ye, C. Fu, Z. Cao, S. An, G. Zheng, and B. Li, “Tracker meets night: A transformer enhancer for uav tracking,” IEEE Robotics and Automation Letters, vol. 7, no. 2, pp. 3866–3873, 2022.
- B. Li, C. Fu, F. Ding, J. Ye, and F. Lin, “Adtrack: Target-aware dual filter learning for real-time anti-dark uav tracking,” in 2021 IEEE international conference on robotics and automation (ICRA). IEEE, 2021, pp. 496–502.
- C. Fu, L. Yao, H. Zuo, G. Zheng, and J. Pan, “Sam-da: Uav tracks anything at night with sam-powered domain adaptation,” arXiv preprint arXiv:2307.01024, 2023.
- A. Kirillov, E. Mintun, N. Ravi, H. Mao, C. Rolland, L. Gustafson, T. Xiao, S. Whitehead, A. C. Berg, W.-Y. Lo et al., “Segment anything,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023, pp. 4015–4026.
- H. Fan, H. A. Miththanthaya, S. R. Rajan, X. Liu, Z. Zou, Y. Lin, H. Ling et al., “Transparent object tracking benchmark,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 10 734–10 743.
- Y. Wu, J. Lim, and M.-H. Yang, “Online object tracking: A benchmark,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2013, pp. 2411–2418.
- ——, “Object tracking benchmark,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 37, pp. 1834–1848, 2015.
- M. Kristan, J. Matas, A. Leonardis, T. Vojíř, R. Pflugfelder, G. Fernandez, G. Nebehay, F. Porikli, and L. Čehovin, “A novel performance evaluation methodology for single-target trackers,” IEEE transactions on pattern analysis and machine intelligence, vol. 38, no. 11, pp. 2137–2155, 2016.
- A. Li, M. Lin, Y. Wu, M.-H. Yang, and S. Yan, “Nus-pro: A new visual tracking challenge,” IEEE transactions on pattern analysis and machine intelligence, vol. 38, no. 2, pp. 335–349, 2015.
- H. Fan, H. Bai, L. Lin, F. Yang, P. Chu, G. Deng, S. Yu, Harshit, M. Huang, J. Liu et al., “Lasot: A high-quality large-scale single object tracking benchmark,” International Journal of Computer Vision, vol. 129, pp. 439–461, 2021.
- M. Muller, A. Bibi, S. Giancola, S. Alsubaihi, and B. Ghanem, “Trackingnet: A large-scale dataset and benchmark for object tracking in the wild,” in Proceedings of the European conference on computer vision (ECCV), 2018, pp. 300–317.
- L. Huang, X. Zhao, and K. Huang, “Got-10k: A large high-diversity benchmark for generic object tracking in the wild,” IEEE transactions on pattern analysis and machine intelligence, vol. 43, no. 5, pp. 1562–1577, 2019.
- M. Mueller, N. G. Smith, and B. Ghanem, “A benchmark and simulator for uav tracking,” in European Conference on Computer Vision, 2016.
- S. Li and D.-Y. Yeung, “Visual object tracking for unmanned aerial vehicles: A benchmark and new motion models,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 31, no. 1, 2017.
- C. Li, X. Liang, Y. Lu, N. Zhao, and J. Tang, “Rgb-t object tracking: Benchmark and baseline,” Pattern Recognition, vol. 96, p. 106977, 2019.
- H. Huang, Y. Xu, Y. Chen, and S.-K. Yeung, “360vot: A new benchmark dataset for omnidirectional visual object tracking,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023, pp. 20 566–20 576.
- X. Wang, K. Ma, Q. Liu, Y. Zou, and Y. Fu, “Multi-object tracking in the dark,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2024, pp. 382–392.
- L. Ma, T. Ma, R. Liu, X. Fan, and Z. Luo, “Toward fast, flexible, and robust low-light image enhancement,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2022, pp. 5637–5646.
- J. Hou, Z. Zhu, J. Hou, H. Liu, H. Zeng, and H. Yuan, “Global structure-aware diffusion process for low-light image enhancement,” Advances in Neural Information Processing Systems, vol. 36, 2024.
- W. Wu, J. Weng, P. Zhang, X. Wang, W. Yang, and J. Jiang, “Uretinex-net: Retinex-based deep unfolding network for low-light image enhancement,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2022, pp. 5901–5910.
- X. Wei, Y. Bai, Y. Zheng, D. Shi, and Y. Gong, “Autoregressive visual tracking,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 9697–9706.
- Q. Wu, T. Yang, Z. Liu, B. Wu, Y. Shan, and A. B. Chan, “Dropmae: Masked autoencoders with spatial-attention dropout for tracking tasks,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 14 561–14 571.
- J. Xie, B. Zhong, Z. Mo, S. Zhang, L. Shi, S. Song, and R. Ji, “Autoregressive queries for adaptive tracking with spatio-temporal transformers,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2024, pp. 19 300–19 309.
- Y. Cai, J. Liu, J. Tang, and G. Wu, “Robust object modeling for visual tracking,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023, pp. 9589–9600.
- S. Gao, C. Zhou, and J. Zhang, “Generalized relation modeling for transformer tracking,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 18 686–18 695.
- B. Yan, H. Peng, J. Fu, D. Wang, and H. Lu, “Learning spatio-temporal transformer for visual tracking,” in Proceedings of the IEEE/CVF international conference on computer vision, 2021, pp. 10 448–10 457.
- X. Chen, H. Peng, D. Wang, H. Lu, and H. Hu, “Seqtrack: Sequence to sequence learning for visual object tracking,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2023, pp. 14 572–14 581.
- Y. Kou, J. Gao, B. Li, G. Wang, W. Hu, Y. Wang, and L. Li, “Zoomtrack: target-aware non-uniform resizing for efficient visual tracking,” Advances in Neural Information Processing Systems, vol. 36, 2024.
- B. Chen, P. Li, L. Bai, L. Qiao, Q. Shen, B. Li, W. Gan, W. Wu, and W. Ouyang, “Backbone is all your need: A simplified architecture for visual object tracking,” in European Conference on Computer Vision. Springer, 2022, pp. 375–392.
- Z. Song, J. Yu, Y.-P. P. Chen, and W. Yang, “Transformer tracking with cyclic shifting window attention,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2022, pp. 8791–8800.
- F. Li, C. Tian, W. Zuo, L. Zhang, and M.-H. Yang, “Learning spatial-temporal regularized correlation filters for visual tracking,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 4904–4913.
- M. Danelljan, G. Hager, F. Shahbaz Khan, and M. Felsberg, “Adaptive decontamination of the training set: A unified formulation for discriminative visual tracking,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 1430–1438.
- ——, “Learning spatially regularized correlation filters for visual tracking,” in Proceedings of the IEEE international conference on computer vision, 2015, pp. 4310–4318.
- H. Kiani Galoogahi, A. Fagg, and S. Lucey, “Learning background-aware correlation filters for visual tracking,” in Proceedings of the IEEE international conference on computer vision, 2017, pp. 1135–1143.
- M. Danelljan, G. Bhat, F. Shahbaz Khan, and M. Felsberg, “Eco: Efficient convolution operators for tracking,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 6638–6646.
- M. Mueller, N. Smith, and B. Ghanem, “Context-aware correlation filter tracking,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 1396–1404.
- L. Bertinetto, J. Valmadre, S. Golodetz, O. Miksik, and P. H. Torr, “Staple: Complementary learners for real-time tracking,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 1401–1409.
- Y. Li, M. Liu, Y. Wu, X. Wang, X. Yang, and S. Li, “Learning adaptive and view-invariant vision transformer for real-time uav tracking,” in Forty-first International Conference on Machine Learning, 2024.
- M. Danelljan, G. Häger, F. Khan, and M. Felsberg, “Accurate scale estimation for robust visual tracking,” in British machine vision conference, Nottingham, September 1-5, 2014. Bmva Press, 2014.
- C. Wang, L. Zhang, L. Xie, and J. Yuan, “Kernel cross-correlator,” in Proceedings of the AAAI conference on artificial intelligence, vol. 32, no. 1, 2018.
- Y. Li and J. Zhu, “A scale adaptive kernel correlation filter tracker with feature integration,” in Computer Vision-ECCV 2014 Workshops: Zurich, Switzerland, September 6-7 and 12, 2014, Proceedings, Part II 13. Springer, 2015, pp. 254–265.
- P. Blatter, M. Kanakis, M. Danelljan, and L. Van Gool, “Efficient visual tracking with exemplar transformers,” in Proceedings of the IEEE/CVF winter conference on applications of computer vision, 2023, pp. 1571–1581.
- M. Danelljan, G. Häger, F. S. Khan, and M. Felsberg, “Discriminative scale space tracking,” IEEE transactions on pattern analysis and machine intelligence, vol. 39, no. 8, pp. 1561–1575, 2016.
- L. Zhou, Z. Zhou, K. Mao, and Z. He, “Joint visual grounding and tracking with natural language specification,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2023, pp. 23 151–23 160.
- J. F. Henriques, R. Caseiro, P. Martins, and J. Batista, “High-speed tracking with kernelized correlation filters,” IEEE transactions on pattern analysis and machine intelligence, vol. 37, no. 3, pp. 583–596, 2014.
- B. Babenko, M.-H. Yang, and S. Belongie, “Robust object tracking with online multiple instance learning,” IEEE transactions on pattern analysis and machine intelligence, vol. 33, no. 8, pp. 1619–1632, 2010.