A Real-Time Framework for Domain-Adaptive Underwater Object Detection with Image Enhancement (2403.19079v1)
Abstract: In recent years, significant progress has been made in the field of underwater image enhancement (UIE). However, its practical utility for high-level vision tasks, such as underwater object detection (UOD) in Autonomous Underwater Vehicles (AUVs), remains relatively unexplored. It may be attributed to several factors: (1) Existing methods typically employ UIE as a pre-processing step, which inevitably introduces considerable computational overhead and latency. (2) The process of enhancing images prior to training object detectors may not necessarily yield performance improvements. (3) The complex underwater environments can induce significant domain shifts across different scenarios, seriously deteriorating the UOD performance. To address these challenges, we introduce EnYOLO, an integrated real-time framework designed for simultaneous UIE and UOD with domain-adaptation capability. Specifically, both the UIE and UOD task heads share the same network backbone and utilize a lightweight design. Furthermore, to ensure balanced training for both tasks, we present a multi-stage training strategy aimed at consistently enhancing their performance. Additionally, we propose a novel domain-adaptation strategy to align feature embeddings originating from diverse underwater environments. Comprehensive experiments demonstrate that our framework not only achieves state-of-the-art (SOTA) performance in both UIE and UOD tasks, but also shows superior adaptability when applied to different underwater scenarios. Our efficiency analysis further highlights the substantial potential of our framework for onboard deployment.
- A. Jesus, C. Zito, C. Tortorici, E. Roura, and G. De Masi, “Underwater object classification and detection: first results and open challenges,” OCEANS 2022-Chennai, pp. 1–6, 2022.
- K. Katija, “Autonomous agents for observing marine life,” Science Robotics, vol. 8, no. 80, p. eadi6428, 2023.
- D. Akkaynak and T. Treibitz, “Sea-thru: A method for removing water from underwater images,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 1682–1691, 2019.
- J. Wen, J. Cui, Z. Zhao, R. Yan, Z. Gao, L. Dou, and B. M. Chen, “Syreanet: A physically guided underwater image enhancement framework integrating synthetic and real images,” in 2023 IEEE International Conference on Robotics and Automation (ICRA), pp. 5177–5183, 2023.
- L. Jiang, Y. Wang, Q. Jia, S. Xu, Y. Liu, X. Fan, H. Li, R. Liu, X. Xue, and R. Wang, “Underwater species detection using channel sharpening attention,” in Proceedings of the 29th ACM International Conference on Multimedia, pp. 4259–4267, 2021.
- F. Zocco, C.-I. Huang, H.-C. Wang, M. O. Khyam, and M. Van, “Towards more efficient efficientdets and low-light real-time marine debris detection,” ArXiv, vol. abs/2203.07155, 2022.
- S. Sun, W. Ren, T. Wang, and X. Cao, “Rethinking image restoration for object detection,” Advances in Neural Information Processing Systems, vol. 35, pp. 4461–4474, 2022.
- W. Liu, G. Ren, R. Yu, S. Guo, J. Zhu, and L. Zhang, “Image-adaptive yolo for object detection in adverse weather conditions,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 36, pp. 1792–1800, 2022.
- Y.-W. Chen and S.-C. Pei, “Domain adaptation for underwater image enhancement via content and style separation,” arXiv preprint arXiv:2202.08537, 2022.
- P. Drews, E. Nascimento, F. Moraes, S. Botelho, and M. Campos, “Transmission estimation in underwater single images,” in Proceedings of the IEEE international conference on computer vision workshops, pp. 825–830, 2013.
- D. Berman, T. Treibitz, and S. Avidan, “Diving into haze-lines: Color restoration of underwater images,” in Proc. British Machine Vision Conference (BMVC), vol. 1, 2017.
- R. Wang, Y. Zhang, and J. Zhang, “An efficient swin transformer-based method for underwater image enhancement,” Multimedia Tools and Applications, vol. 82, no. 12, pp. 18691–18708, 2023.
- Z. Liu, Y. Lin, Y. Cao, H. Hu, Y. Wei, Z. Zhang, S. Lin, and B. Guo, “Swin transformer: Hierarchical vision transformer using shifted windows,” in Proceedings of the IEEE/CVF international conference on computer vision, pp. 10012–10022, 2021.
- S. Huang, K. Wang, H. Liu, J. Chen, and Y. Li, “Contrastive semi-supervised learning for underwater image restoration via reliable bank,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 18145–18155, 2023.
- S. Jamieson, J. P. How, and Y. Girdhar, “Deepseecolor: Realtime adaptive color correction for autonomous underwater vehicles via deep learning methods,” in 2023 IEEE International Conference on Robotics and Automation (ICRA), pp. 3095–3101, 2023.
- S. Ren, K. He, R. Girshick, and J. Sun, “Faster r-cnn: Towards real-time object detection with region proposal networks,” Advances in neural information processing systems, vol. 28, 2015.
- J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, “You only look once: Unified, real-time object detection,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 779–788, 2016.
- M. A. Syariz, C.-H. Lin, M. V. Nguyen, L. M. Jaelani, and A. C. Blanco, “Waternet: A convolutional neural network for chlorophyll-a concentration retrieval,” Remote Sensing, vol. 12, no. 12, p. 1966, 2020.
- B. Fan, W. Chen, Y. Cong, and J. Tian, “Dual refinement underwater object detection network,” in Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XX 16, pp. 275–291, Springer, 2020.
- C. Li, H. Zhou, Y. Liu, C. Yang, Y. Xie, Z. Li, and L. Zhu, “Detection-friendly dehazing: Object detection in real-world hazy scenes,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 2023.
- N. Cheng, H. Xie, X. Zhu, and H. Wang, “Joint image enhancement learning for marine object detection in natural scene,” Engineering Applications of Artificial Intelligence, vol. 120, p. 105905, 2023.
- X. Liu, Z. Gao, and B. M. Chen, “Ipmgan: Integrating physical model and generative adversarial network for underwater image enhancement,” Neurocomputing, vol. 453, pp. 538–551, 2021.
- Y.-J. Li, X. Dai, C.-Y. Ma, Y.-C. Liu, K. Chen, B. Wu, Z. He, K. Kitani, and P. Vajda, “Cross-domain adaptive teacher for object detection,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 7581–7590, 2022.
- P. M. Uplavikar, Z. Wu, and Z. Wang, “All-in-one underwater image enhancement using domain-adversarial learning.,” in CVPR workshops, pp. 1–8, 2019.
- H. Liu, P. Song, and R. Ding, “Towards domain generalization in underwater object detection,” in 2020 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), pp. 1971–1975, IEEE, 2020.
- J. Yoo, Y. Uh, S. Chun, B. Kang, and J.-W. Ha, “Photorealistic style transfer via wavelet transforms,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 9036–9045, 2019.
- K. He, G. Gkioxari, P. Dollar, and R. Girshick, “Mask r-cnn,” in Proceedings of the IEEE International Conference on Computer Vision (ICCV), Oct 2017.
- D. Wang, J. Wen, Y. Wang, X. Huang, and F. Pei, “End-to-end self-driving using deep neural networks with multi-auxiliary tasks,” Automotive Innovation, vol. 2, pp. 127–136, 2019.
- G. Jocher, “YOLOv5 by Ultralytics,” May 2020.
- C.-Y. Wang, H.-Y. M. Liao, Y.-H. Wu, P.-Y. Chen, J.-W. Hsieh, and I.-H. Yeh, “Cspnet: A new backbone that can enhance learning capability of cnn,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition workshops, pp. 390–391, 2020.
- B. Sun and K. Saenko, “Deep coral: Correlation alignment for deep domain adaptation,” in Computer Vision–ECCV 2016 Workshops: Amsterdam, The Netherlands, October 8-10 and 15-16, 2016, Proceedings, Part III 14, pp. 443–450, Springer, 2016.
- C. Liu, H. Li, S. Wang, M. Zhu, D. Wang, X. Fan, and Z. Wang, “A dataset and benchmark of underwater object detection for robot picking,” in 2021 IEEE International Conference on Multimedia & Expo Workshops (ICMEW), pp. 1–6, IEEE, 2021.
- A. Bochkovskiy, C.-Y. Wang, and H.-Y. M. Liao, “Yolov4: Optimal speed and accuracy of object detection,” arXiv preprint arXiv:2004.10934, 2020.
- C. Li, C. Guo, W. Ren, R. Cong, J. Hou, S. Kwong, and D. Tao, “An underwater image enhancement benchmark dataset and beyond,” IEEE Transactions on Image Processing, vol. 29, pp. 4376–4389, 2019.
- T. Falk, D. Mai, R. Bensch, Ö. Çiçek, A. Abdulkadir, Y. Marrakchi, A. Böhm, J. Deubner, Z. Jäckel, K. Seiwald, et al., “U-net: deep learning for cell counting, detection, and morphometry,” Nature methods, vol. 16, no. 1, pp. 67–70, 2019.
- X. Fu and X. Cao, “Underwater image enhancement with global–local networks and compressed-histogram equalization,” Signal Processing: Image Communication, vol. 86, p. 115892, 2020.
- K. Panetta, C. Gao, and S. Agaian, “Human-visual-system-inspired underwater image quality measures,” IEEE Journal of Oceanic Engineering, vol. 41, no. 3, pp. 541–551, 2015.
- M. Yang and A. Sowmya, “An underwater color image quality evaluation metric,” IEEE Transactions on Image Processing, vol. 24, no. 12, pp. 6062–6071, 2015.
- Q. Qin, K. Chang, M. Huang, and G. Li, “Denet: Detection-driven enhancement network for object detection under adverse weather conditions,” in Proceedings of the Asian Conference on Computer Vision, pp. 2813–2829, 2022.