Boosting Visual Recognition in Real-world Degradations via Unsupervised Feature Enhancement Module with Deep Channel Prior (2404.01703v2)
Abstract: The environmental perception of autonomous vehicles in normal conditions have achieved considerable success in the past decade. However, various unfavourable conditions such as fog, low-light, and motion blur will degrade image quality and pose tremendous threats to the safety of autonomous driving. That is, when applied to degraded images, state-of-the-art visual models often suffer performance decline due to the feature content loss and artifact interference caused by statistical and structural properties disruption of captured images. To address this problem, this work proposes a novel Deep Channel Prior (DCP) for degraded visual recognition. Specifically, we observe that, in the deep representation space of pre-trained models, the channel correlations of degraded features with the same degradation type have uniform distribution even if they have different content and semantics, which can facilitate the mapping relationship learning between degraded and clear representations in high-sparsity feature space. Based on this, a novel plug-and-play Unsupervised Feature Enhancement Module (UFEM) is proposed to achieve unsupervised feature correction, where the multi-adversarial mechanism is introduced in the first stage of UFEM to achieve the latent content restoration and artifact removal in high-sparsity feature space. Then, the generated features are transferred to the second stage for global correlation modulation under the guidance of DCP to obtain high-quality and recognition-friendly features. Evaluations of three tasks and eight benchmark datasets demonstrate that our proposed method can comprehensively improve the performance of pre-trained models in real degradation conditions. The source code is available at https://github.com/liyuhang166/Deep_Channel_Prior
- F.-Y. Wang, “A new phase of ieee transactions on intelligent vehicles: Being smart, becoming active, and believing intelligent vehicles,” IEEE Transactions on Intelligent Vehicles, vol. 8, no. 1, pp. 3–15, 2023.
- S. Teng, X. Hu, P. Deng, B. Li, Y. Li, Y. Ai, D. Yang, L. Li, Z. Xuanyuan, F. Zhu et al., “Motion planning for autonomous driving: The state of the art and future perspectives,” IEEE Transactions on Intelligent Vehicles, 2023.
- S. Teng, L. Chen, Y. Ai, Y. Zhou, Z. Xuanyuan, and X. Hu, “Hierarchical interpretable imitation learning for end-to-end autonomous driving,” IEEE Transactions on Intelligent Vehicles, vol. 8, no. 1, pp. 673–683, 2022.
- Z. Liu, J. Cheng, J. Fan, S. Lin, Y. Wang, and X. Zhao, “Multi-modal fusion based on depth adaptive mechanism for 3d object detection,” IEEE Transactions on Multimedia, 2023.
- C. Li, Z. Liu, N. Yang, W. Li, and X. Zhao, “Regional attention network with data-driven modal representation for multimodal trajectory prediction,” Expert Systems with Applications, vol. 232, p. 120808, 2023.
- C. Li, Z. Liu, S. Lin, Y. Wang, and X. Zhao, “Intention-convolution and hybrid-attention network for vehicle trajectory prediction,” Expert Systems with Applications, vol. 236, p. 121412, 2024.
- Z. Liu, M. Qi, C. Shen, Y. Fang, and X. Zhao, “Cascade saccade machine learning network with hierarchical classes for traffic sign detection,” Sustainable Cities and Society, vol. 67, p. 102700, 2021.
- Z. Liu, C. Shen, X. Fan, G. Zeng, and X. Zhao, “Scale-aware limited deformable convolutional neural networks for traffic sign detection and classification,” IET Intelligent Transport Systems, vol. 14, no. 12, pp. 1712–1722, 2020.
- Y. Wang, L. Peng, L. Li, Y. Cao, and Z.-J. Zha, “Decoupling-and-aggregating for image exposure correction,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 18 115–18 124.
- H. Wang, L. Peng, Y. Sun, Z. Wan, Y. Wang, and Y. Cao, “Brightness perceiving for recursive low-light image enhancement,” IEEE Transactions on Artificial Intelligence, 2023.
- Y. Wang, Y. Cao, J. Zhang, F. Wu, and Z.-J. Zha, “Leveraging deep statistics for underwater image enhancement,” ACM Transactions on Multimedia Computing, Communications, and Applications (TOMM), vol. 17, no. 3s, pp. 1–20, 2021.
- I. Kim, S. Han, J.-w. Baek, S.-J. Park, J.-J. Han, and J. Shin, “Quality-agnostic image recognition via invertible decoder,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 12 257–12 266.
- T. Son, J. Kang, N. Kim, S. Cho, and S. Kwak, “Urie: Universal image enhancement for visual recognition in the wild,” in Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part IX 16. Springer, 2020, pp. 749–765.
- Y. Wang, Y. Cao, Z.-J. Zha, J. Zhang, and Z. Xiong, “Deep degradation prior for low-quality image classification,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 11 049–11 058.
- Z. Yang, W. Dong, X. Li, J. Wu, L. Li, and G. Shi, “Self-feature distillation with uncertainty modeling for degraded image recognition,” in European Conference on Computer Vision. Springer, 2022, pp. 552–569.
- Z. Liu, N. Yang, Y. Wang, Y. Li, X. Zhao, and F.-Y. Wang, “Enhancing traffic object detection in variable illumination with rgb-event fusion,” arXiv preprint arXiv:2311.00436, 2023.
- N. Yang, Z. Liu, S. Ma, Y. Sun, Y. He, and Y. Wang, “Joint intensity and event framework for vehicle detection in degraded conditions,” in 2023 7th International Conference on Transportation Information and Safety (ICTIS). IEEE, 2023, pp. 1568–1574.
- Y. Pei, Y. Huang, Q. Zou, Y. Lu, and S. Wang, “Does haze removal help cnn-based image classification?” in Proceedings of the European Conference on Computer Vision (ECCV), 2018, pp. 682–697.
- W. Wang, F. Wen, Z. Yan, and P. Liu, “Optimal transport for unsupervised denoising learning,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 45, no. 2, pp. 2104–2118, 2022.
- S. Zhao, L. Zhang, Y. Shen, and Y. Zhou, “Refinednet: A weakly supervised refinement framework for single image dehazing,” IEEE Transactions on Image Processing, vol. 30, pp. 3391–3404, 2021.
- S. Zhao, Z. Zhang, R. Hong, M. Xu, Y. Yang, and M. Wang, “Fcl-gan: A lightweight and real-time baseline for unsupervised blind image deblurring,” in Proceedings of the 30th ACM International Conference on Multimedia, 2022, pp. 6220–6229.
- Y. Wang, J. Zhang, Y. Cao, and Z. Wang, “A deep cnn method for underwater image enhancement,” in 2017 IEEE international conference on image processing (ICIP). IEEE, 2017, pp. 1382–1386.
- Z. Du, J. Li, H. Su, L. Zhu, and K. Lu, “Cross-domain gradient discrepancy minimization for unsupervised domain adaptation,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2021, pp. 3937–3946.
- M. Long, Y. Cao, J. Wang, and M. Jordan, “Learning transferable features with deep adaptation networks,” in International conference on machine learning. PMLR, 2015, pp. 97–105.
- E. Tzeng, J. Hoffman, K. Saenko, and T. Darrell, “Adversarial discriminative domain adaptation,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 7167–7176.
- W. Zhang, W. Ouyang, W. Li, and D. Xu, “Collaborative and adversarial network for unsupervised domain adaptation,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 3801–3809.
- Y. P. Loh and C. S. Chan, “Getting to know low-light images with the exclusively dark dataset,” Computer Vision and Image Understanding, vol. 178, pp. 30–42, 2019.
- D. Hendrycks and T. Dietterich, “Benchmarking neural network robustness to common corruptions and perturbations,” arXiv preprint arXiv:1903.12261, 2019.
- Y. Kim and J. Shin, “Efficient and robust object detection against multi-type corruption using complete edge based on lightweight parameter isolation,” IEEE Transactions on Intelligent Vehicles, 2024.
- E. D. Cubuk, B. Zoph, D. Mane, V. Vasudevan, and Q. V. Le, “Autoaugment: Learning augmentation policies from data,” arXiv preprint arXiv:1805.09501, 2018.
- D. Hendrycks, N. Mu, E. D. Cubuk, B. Zoph, J. Gilmer, and B. Lakshminarayanan, “Augmix: A simple data processing method to improve robustness and uncertainty,” arXiv preprint arXiv:1912.02781, 2019.
- C. Xie, M. Tan, B. Gong, J. Wang, A. L. Yuille, and Q. V. Le, “Adversarial examples improve image recognition,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 819–828.
- Y. Jiang, X. Gong, D. Liu, Y. Cheng, C. Fang, X. Shen, J. Yang, P. Zhou, and Z. Wang, “Enlightengan: Deep light enhancement without paired supervision,” IEEE transactions on image processing, vol. 30, pp. 2340–2349, 2021.
- S. Zhao, Z. Zhang, R. Hong, M. Xu, H. Zhang, M. Wang, and S. Yan, “Crnet: Unsupervised color retention network for blind motion deblurring,” in Proceedings of the 30th ACM International Conference on Multimedia, 2022, pp. 6193–6201.
- Y. Liang, J. Fan, X. Zheng, Y. Wang, L. Huangfu, V. Ghavate, and Z. Yu, “An interpretable image denoising framework via dual disentangled representation learning,” IEEE Transactions on Intelligent Vehicles, 2023.
- R. W. Liu, Y. Lu, Y. Guo, W. Ren, F. Zhu, and Y. Lv, “Aioenet: All-in-one low-visibility enhancement to improve visual perception for intelligent marine vehicles under severe weather conditions,” IEEE Transactions on Intelligent Vehicles, 2023.
- W. Zhou, R. Zhang, L. Li, G. Yue, J. Gong, H. Chen, and H. Liu, “Dehazed image quality evaluation: From partial discrepancy to blind perception,” IEEE Transactions on Intelligent Vehicles, 2024.
- C. Guo, C. Li, J. Guo, C. C. Loy, J. Hou, S. Kwong, and R. Cong, “Zero-reference deep curve estimation for low-light image enhancement,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 1780–1789.
- Q. Jiang, Y. Mao, R. Cong, W. Ren, C. Huang, and F. Shao, “Unsupervised decomposition and correction network for low-light image enhancement,” IEEE Transactions on Intelligent Transportation Systems, vol. 23, no. 10, pp. 19 440–19 455, 2022.
- C. Li, C. Guo, and C. C. Loy, “Learning to enhance low-light image via zero-reference deep curve estimation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 44, no. 8, pp. 4225–4238, 2021.
- J.-Y. Zhu, T. Park, P. Isola, and A. A. Efros, “Unpaired image-to-image translation using cycle-consistent adversarial networks,” in Proceedings of the IEEE international conference on computer vision, 2017, pp. 2223–2232.
- B. Li, Y. Gou, S. Gu, J. Z. Liu, J. T. Zhou, and X. Peng, “You only look yourself: Unsupervised and untrained single image dehazing neural network,” International Journal of Computer Vision, vol. 129, pp. 1754–1767, 2021.
- R. Neshatavar, M. Yavartanoo, S. Son, and K. M. Lee, “Cvf-sid: Cyclic multi-variate function for self-supervised image denoising by disentangling noise from image,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 17 583–17 591.
- B. Lu, J.-C. Chen, and R. Chellappa, “Uid-gan: Unsupervised image deblurring via disentangled representations,” IEEE Transactions on Biometrics, Behavior, and Identity Science, vol. 2, no. 1, pp. 26–39, 2019.
- Y. Fu, Y. Hong, L. Chen, and S. You, “Le-gan: Unsupervised low-light image enhancement network using attention module and identity invariant loss,” Knowledge-Based Systems, vol. 240, p. 108010, 2022.
- K. Saito, K. Watanabe, Y. Ushiku, and T. Harada, “Maximum classifier discrepancy for unsupervised domain adaptation,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 3723–3732.
- M. Caron, P. Bojanowski, A. Joulin, and M. Douze, “Deep clustering for unsupervised learning of visual features,” in Proceedings of the European conference on computer vision (ECCV), 2018, pp. 132–149.
- Z. Han, H. Sun, and Y. Yin, “Learning transferable parameters for unsupervised domain adaptation,” IEEE Transactions on Image Processing, vol. 31, pp. 6424–6439, 2022.
- B. Li, W. Ren, D. Fu, D. Tao, D. Feng, W. Zeng, and Z. Wang, “Benchmarking single-image dehazing and beyond,” IEEE Transactions on Image Processing, vol. 28, no. 1, pp. 492–505, 2018.
- L. Van der Maaten and G. Hinton, “Visualizing data using t-sne.” Journal of machine learning research, vol. 9, no. 11, 2008.
- L. Wang, V. Sindagi, and V. Patel, “High-quality facial photo-sketch synthesis using multi-adversarial networks,” in 2018 13th IEEE international conference on automatic face & gesture recognition (FG 2018). IEEE, 2018, pp. 83–90.
- I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial nets,” Advances in neural information processing systems, vol. 27, 2014.
- K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” arXiv preprint arXiv:1409.1556, 2014.
- K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778.
- J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, “You only look once: Unified, real-time object detection,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 779–788.
- L.-C. Chen, Y. Zhu, G. Papandreou, F. Schroff, and H. Adam, “Encoder-decoder with atrous separable convolution for semantic image segmentation,” in Proceedings of the European conference on computer vision (ECCV), 2018, pp. 801–818.
- M. A. Kenk and M. Hassaballah, “Dawn: vehicle detection in adverse weather nature dataset,” arXiv preprint arXiv:2008.05402, 2020.
- C. Sakaridis, D. Dai, and L. Van Gool, “Acdc: The adverse conditions dataset with correspondences for semantic driving scene understanding,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 10 765–10 775.
- C. Sakaridis, D. Dai, and L. V. Gool, “Guided curriculum model adaptation and uncertainty-aware evaluation for semantic nighttime image segmentation,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019, pp. 7374–7383.
- D. Dai and L. Van Gool, “Dark model adaptation: Semantic image segmentation from daytime to nighttime,” in 2018 21st International Conference on Intelligent Transportation Systems (ITSC). IEEE, 2018, pp. 3819–3824.
- J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei, “Imagenet: A large-scale hierarchical image database,” in 2009 IEEE conference on computer vision and pattern recognition. Ieee, 2009, pp. 248–255.
- M. Everingham, L. Van Gool, C. K. Williams, J. Winn, and A. Zisserman, “The pascal visual object classes (voc) challenge,” International journal of computer vision, vol. 88, pp. 303–338, 2010.
- T.-Y. Lin, M. Maire, S. Belongie, J. Hays, P. Perona, D. Ramanan, P. Dollár, and C. L. Zitnick, “Microsoft coco: Common objects in context,” in Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part V 13. Springer, 2014, pp. 740–755.
- M. Cordts, M. Omran, S. Ramos, T. Rehfeld, M. Enzweiler, R. Benenson, U. Franke, S. Roth, and B. Schiele, “The cityscapes dataset for semantic urban scene understanding,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 3213–3223.
- Y. Wang, X. Yan, D. Guan, M. Wei, Y. Chen, X.-P. Zhang, and J. Li, “Cycle-snspgan: Towards real-world image dehazing via cycle spectral normalized soft likelihood estimation patch gan,” IEEE Transactions on Intelligent Transportation Systems, vol. 23, no. 11, pp. 20 368–20 382, 2022.
- Y. Yang, C. Wang, R. Liu, L. Zhang, X. Guo, and D. Tao, “Self-augmented unpaired image dehazing via density and depth decomposition,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2022, pp. 2037–2046.
- P. Ling, H. Chen, X. Tan, Y. Jin, and E. Chen, “Single image dehazing using saturation line prior,” IEEE Transactions on Image Processing, 2023.
- Y. Zheng, J. Zhan, S. He, J. Dong, and Y. Du, “Curricular contrastive regularization for physics-aware single image dehazing,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 5785–5794.
- K. Zhang, W. Luo, Y. Zhong, L. Ma, B. Stenger, W. Liu, and H. Li, “Deblurring by realistic blurring,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 2737–2746.
- S. Yang, M. Ding, Y. Wu, Z. Li, and J. Zhang, “Implicit neural representation for cooperative low-light image enhancement,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023, pp. 12 918–12 927.
- N. Zheng, M. Zhou, Y. Dong, X. Rui, J. Huang, C. Li, and F. Zhao, “Empowering low-light image enhancer through customized learnable priors,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023, pp. 12 559–12 569.
- R. R. Selvaraju, M. Cogswell, A. Das, R. Vedantam, D. Parikh, and D. Batra, “Grad-cam: Visual explanations from deep networks via gradient-based localization,” in Proceedings of the IEEE international conference on computer vision, 2017, pp. 618–626.
- S. Dodge and L. Karam, “Understanding how image quality affects deep neural networks,” in 2016 eighth international conference on quality of multimedia experience (QoMEX). IEEE, 2016, pp. 1–6.
- M. D. Zeiler and R. Fergus, “Visualizing and understanding convolutional networks,” in Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part I 13. Springer, 2014, pp. 818–833.
- R. Belaroussi and D. Gruyer, “Impact of reduced visibility from fog on traffic sign detection,” in 2014 IEEE intelligent vehicles symposium proceedings. IEEE, 2014, pp. 1302–1306.
- M. Maaz, A. Shaker, H. Cholakkal, S. Khan, S. W. Zamir, R. M. Anwer, and F. Shahbaz Khan, “Edgenext: efficiently amalgamated cnn-transformer architecture for mobile vision applications,” in European Conference on Computer Vision. Springer, 2022, pp. 3–20.