Contrastive Learning for Compact Single Image Dehazing
The paper "Contrastive Learning for Compact Single Image Dehazing" introduces a novel approach to tackle the ill-posed problem of single image dehazing. Traditional deep learning methods for dehazing typically rely on clear images as positive samples to guide neural network training, but they fail to leverage negative samples, namely hazy images, which could provide valuable information for improving the dehazing process. Furthermore, many existing models address dehazing by significantly increasing the network's depth and width, resulting in high computational and memory demands. This research proposes a new methodology employing contrastive regularization (CR) to make more efficient use of both positive and negative samples in representation learning.
CR is built upon contrastive learning principles, ensuring that the outputs from the dehazing network are closer to the clear images while being furthest from hazy images in the representation space. This approach not only enhances dehazing performance but also reduces artifacts and color distortions commonly seen in methods relying solely on reconstruction loss.
The authors introduce a compact yet effective autoencoder-like (AE) network architecture, AECR-Net, specifically designed to balance performance with computational efficiency. Key components of AECR-Net include an adaptive mixup operation that enhances feature preservation across the network and a dynamic feature enhancement (DFE) module based on deformable convolution that dynamically extends the receptive field for capturing more robust spatial features. These additions equip the network with increased transformation capabilities without excessive resource requirements.
Experimental results demonstrate that AECR-Net outperforms state-of-the-art dehazing techniques on both synthetic and real-world datasets. Notably, AECR-Net achieves a PSNR of 37.17 and an SSIM of 0.9901 on the SOTS synthetic dataset with fewer parameters than many competitors. On real-world datasets like Dense-Haze and NH-HAZE, the proposed method maintains superior performance, reinforcing its robustness and applicability in diverse scenarios.
This paper's implication is twofold: practically, it presents a highly efficient model suitable for deployment in resource-constrained environments, such as mobile devices; theoretically, it paves the way for broader applications of contrastive learning in low-level vision tasks beyond traditional high-level applications.
Future work could explore the integration of more sophisticated representation learning frameworks and evaluate the generalizability of CR and AE-like architectures for other image restoration tasks. Additionally, understanding the impact of contrastive sample selection and its adaptation to changing environmental conditions could further refine dehazing performance. Thus, this paper contributes significantly to the ongoing development of efficient and effective methods for image restoration and enhancement.