Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
184 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

An Underwater Image Enhancement Benchmark Dataset and Beyond (1901.05495v2)

Published 11 Jan 2019 in cs.CV

Abstract: Underwater image enhancement has been attracting much attention due to its significance in marine engineering and aquatic robotics. Numerous underwater image enhancement algorithms have been proposed in the last few years. However, these algorithms are mainly evaluated using either synthetic datasets or few selected real-world images. It is thus unclear how these algorithms would perform on images acquired in the wild and how we could gauge the progress in the field. To bridge this gap, we present the first comprehensive perceptual study and analysis of underwater image enhancement using large-scale real-world images. In this paper, we construct an Underwater Image Enhancement Benchmark (UIEB) including 950 real-world underwater images, 890 of which have the corresponding reference images. We treat the rest 60 underwater images which cannot obtain satisfactory reference images as challenging data. Using this dataset, we conduct a comprehensive study of the state-of-the-art underwater image enhancement algorithms qualitatively and quantitatively. In addition, we propose an underwater image enhancement network (called Water-Net) trained on this benchmark as a baseline, which indicates the generalization of the proposed UIEB for training Convolutional Neural Networks (CNNs). The benchmark evaluations and the proposed Water-Net demonstrate the performance and limitations of state-of-the-art algorithms, which shed light on future research in underwater image enhancement. The dataset and code are available at https://li-chongyi.github.io/proj_benchmark.html.

Citations (1,074)

Summary

  • The paper introduces the UIEB dataset with 950 real-world underwater images and rigorously paired references to benchmark enhancement methods.
  • The paper evaluates state-of-the-art methods using both full-reference (PSNR, SSIM) and non-reference (UCIQE, UIQM) metrics, highlighting discrepancies with human visual assessment.
  • The paper proposes Water-Net, a CNN model leveraging gated fusion and perceptual loss to consistently outperform traditional enhancement techniques.

An Underwater Image Enhancement Benchmark Dataset and Beyond

The paper "An Underwater Image Enhancement Benchmark Dataset and Beyond" presents a comprehensive paper on the challenging task of underwater image enhancement. This problem is paramount in marine engineering and aquatic robotics due to the unique degradation characteristics of underwater images, such as color casts, reduced visibility, and contrast loss caused by wavelength-dependent absorption and scattering, as well as marine snow.

Dataset Construction

To address the gap in evaluating underwater image enhancement algorithms using real-world data, the authors introduce the Underwater Image Enhancement Benchmark (UIEB) dataset. UIEB contains 950 real-world underwater images, with 890 images paired with corresponding reference images and 60 images classified as challenging data without satisfactory reference images. The authors curated this dataset meticulously, ensuring a wide variety of underwater scenes and degradation characteristics. The reference images were selected through a rigorous pairwise comparison by volunteers, ensuring high-quality ground truth data for performance evaluation.

Experimental Evaluation

The paper provides an exhaustive evaluation of state-of-the-art underwater image enhancement methods using the UIEB dataset. Qualitative and quantitative analyses were conducted on methods including fusion-based, retinex-based, and dark channel prior variants.

Qualitative Analysis

The qualitative analysis categorizes underwater images into different types (e.g., greenish, bluish, downward-looking, forward-looking, low and high backscatter scenes). The visual inspection revealed that while the fusion-based method generally performed well across various scenarios, other methods often introduced artifacts, over-enhancement, or color deviations.

Quantitative Analysis

Quantitative assessments were performed using established full-reference metrics (PSNR, SSIM) and non-reference metrics (UCIQE, UIQM). Results demonstrated that commercial applications like Dive+ often outperformed other methods in terms of both full-reference and non-reference metrics. However, the discrepancy between quantitative scores from UCIQE/UIQM and subjective visual quality indicates the need for more reliable non-reference metrics tuned to human visual perception.

Proposed CNN Model: Water-Net

Leveraging the comprehensive UIEB dataset, the authors proposed a Convolutional Neural Network model named Water-Net. Water-Net employs a gated fusion network architecture designed to learn confidence maps for inputs obtained via white balance, histogram equalization, and gamma correction. The use of a perceptual loss function during training aims to optimize for visually pleasing and realistic enhancement results.

Training and Performance

Water-Net was trained on a subset of the UIEB dataset. Experimental results reveal that Water-Net outperforms conventional methods both on typical underwater images and on more challenging cases. The model's effectiveness was further substantiated by user paper evaluations and standard deviation analyses, indicating robust and consistent enhancement performance.

Implications and Future Research Directions

The introduction of the UIEB dataset sets a new benchmark for evaluating underwater image enhancement methods. It enables the rigorous comparison of different approaches under standardized conditions. The dataset also proves instrumental in training deep learning models, as demonstrated by the superior performance of Water-Net.

The paper identifies significant areas for future work, such as extending the dataset to include more challenging scenarios and underwater videos. Moreover, future research could focus on developing more physically accurate models for underwater image formation and metrics for non-reference image quality assessment that better align with human perception.

Conclusion

This paper provides a substantial contribution to the field of underwater image enhancement by addressing the critical lack of a comprehensive real-world dataset and offering a robust baseline model for future research. The provided benchmarks and evaluations lay the groundwork for subsequent advancements in improving the visual quality of underwater images, which will have significant practical and theoretical implications for various applications in marine science and technology.

Github Logo Streamline Icon: https://streamlinehq.com