Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
167 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Case Study of Image Enhancement Algorithms' Effectiveness of Improving Neural Networks' Performance on Adverse Images (2312.09509v2)

Published 15 Dec 2023 in eess.IV

Abstract: Neural Networks (NNs) have become indispensable for applications of Computer Vision (CV) and their use has been ever-growing. NNs are commonly trained for long periods of time on datasets like ImageNet and COCO that have been carefully created to represent common "real-world" environments. When deployed in the field, such as applications of autonomous vehicles, NNs can encounter adverse scenarios that degrade performance. Using image enhancements algorithms to enhance images before being inferenced on a NN model poses an intriguing alternative to retraining, however, published literature on the effectiveness of this solution is scarce. To fill this knowledge gap, we provide a case study on two popular image enhancement algorithms, Histogram Equalization (HE) and Retinex (RX). We simulate four types of adverse scenarios an autonomous vehicle could encounter, dark, over exposed, foggy, and dark & rainy weather conditions. We evaluate the effectiveness of HE and RX using several well established models:, Resnet, GoogleLeNet, YOLO, and a Vision Transformer.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (25)
  1. T. Wang, M. Sun, and K. Hu, “Dilated deep residual network for image denoising,” in 2017 IEEE 29th international conference on tools with artificial intelligence (ICTAI), pp. 1272–1279, IEEE, 2017.
  2. Y. Nagano and Y. Kikuta, “Srgan for super-resolving low-resolution food images,” in Proceedings of the Joint Workshop on Multimedia for Cooking and Eating Activities and Multimedia Assisted Dietary Management, pp. 33–37, 2018.
  3. H. Murad, A. Al Razi, and R. Rahman, “Colornet: A deep learning-based approach to colorize the historical documentaries of bangladesh,” in Proceedings of International Conference on Fourth Industrial Revolution and Beyond 2021, pp. 3–14, Springer, 2022.
  4. J. Li and P. Fang, “Hdrnet: Single-image-based hdr reconstruction using channel attention cnn,” in Proceedings of the 2019 4th International Conference on Multimedia Systems and Signal Processing, pp. 119–124, 2019.
  5. C. Wei, W. Wang, W. Yang, and J. Liu, “Deep retinex decomposition for low-light enhancement,” arXiv preprint arXiv:1808.04560, 2018.
  6. P. L. Suárez, A. D. Sappa, B. X. Vintimilla, and R. I. Hammoud, “Deep learning based single image dehazing,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 1169–1176, 2018.
  7. H. Wang, Q. Xie, Q. Zhao, and D. Meng, “A model-driven deep neural network for single image rain removal,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 3103–3112, 2020.
  8. X. Chen, “Image enhancement effect on the performance of convolutional neural networks,” 2019.
  9. Y. Qi, Z. Yang, W. Sun, M. Lou, J. Lian, W. Zhao, X. Deng, and Y. Ma, “A comprehensive overview of image enhancement techniques,” Archives of Computational Methods in Engineering, pp. 1–25, 2021.
  10. K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770–778, 2016.
  11. J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, “You only look once: Unified, real-time object detection,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 779–788, 2016.
  12. C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke, and A. Rabinovich, “Going deeper with convolutions,” in 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1–9, 2015.
  13. A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly, et al., “An image is worth 16x16 words: Transformers for image recognition at scale,” arXiv preprint arXiv:2010.11929, 2020.
  14. “vit_b_16.” https://pytorch.org/vision/stable/models/generated/torchvision.models.vit_b_16.html#torchvision.models.vit_b_16.
  15. J.-M. Morel, A.-B. Petro, and C. Sbert, “Fast implementation of color constancy algorithms,” Proceedings of SPIE - The International Society for Optical Engineering, vol. 7241, 01 2009.
  16. “Opencv histogram equalization.” https://docs.opencv.org/3.4/d4/d1b/tutorial_histogram_equalization.html.
  17. Muggledy, “GitHub - muggledy/retinex: Retinex Algorithms: python code for MSRCR, MSRCP..” https://github.com/muggledy/retinex.
  18. “Resnet18.” https://pytorch.org/vision/stable/models/generated/torchvision.models.resnet18.html#torchvision.models.resnet18.
  19. “Resnet50.” https://pytorch.org/vision/stable/models/generated/torchvision.models.resnet50.html#torchvision.models.resnet50.
  20. “Resnet101.” https://pytorch.org/vision/stable/models/generated/torchvision.models.resnet101.html#torchvision.models.resnet101.
  21. “Yolov5n.” https://github.com/ultralytics/yolov5/releases/download/v7.0/yolov5n.pt.
  22. “Yolov5m.” https://github.com/ultralytics/yolov5/releases/download/v7.0/yolov5m.pt.
  23. “Yolov8n.” https://github.com/ultralytics/assets/releases/download/v0.0.0/yolov8n.pt.
  24. “Yolov8m.” https://github.com/ultralytics/assets/releases/download/v0.0.0/yolov8m.pt.
  25. “Googlenet.” https://pytorch.org/vision/stable/models/generated/torchvision.models.googlenet.html#torchvision.models.googlenet.

Summary

We haven't generated a summary for this paper yet.