Papers
Topics
Authors
Recent
2000 character limit reached

Deeply-Recursive Convolutional Network for Image Super-Resolution (1511.04491v2)

Published 14 Nov 2015 in cs.CV and cs.LG

Abstract: We propose an image super-resolution method (SR) using a deeply-recursive convolutional network (DRCN). Our network has a very deep recursive layer (up to 16 recursions). Increasing recursion depth can improve performance without introducing new parameters for additional convolutions. Albeit advantages, learning a DRCN is very hard with a standard gradient descent method due to exploding/vanishing gradients. To ease the difficulty of training, we propose two extensions: recursive-supervision and skip-connection. Our method outperforms previous methods by a large margin.

Citations (2,413)

Summary

  • The paper presents a novel deeply-recursive convolutional network that leverages repeated convolutional layers to enlarge the receptive field without increasing parameters.
  • It implements recursive-supervision and skip-connections to counter gradient issues and efficiently preserve low-resolution details.
  • Extensive experiments show superior PSNR and SSIM results on standard datasets like Set5 and Urban100, setting a new state-of-the-art.

Deeply-Recursive Convolutional Network for Image Super-Resolution

Overview

The paper "Deeply-Recursive Convolutional Network for Image Super-Resolution" by Jiwon Kim, Jung Kwon Lee, and Kyoung Mu Lee presents an innovative method for addressing the challenge of single-image super-resolution (SR). The authors utilize a deeply-recursive convolutional network (DRCN) to exploit large image contexts without significantly increasing the number of parameters. Their approach involves recurrently applying the same convolutional layer, which allows for a large receptive field of up to 41x41 pixels, thus enabling the network to capture long-range pixel dependencies crucial for image restoration tasks.

Key Contributions

Deeply-Recursive Convolutional Network

The core idea in this paper is the deeply-recursive application of convolutional layers. The authors argue that increasing recursion depth improves SR performance without adding new parameters, thus avoiding overfitting and high memory utilization. They demonstrate this by recursively applying the same convolutional layer up to 16 times, significantly widening the receptive field and boosting performance compared to previous methods such as SRCNN.

Recursive-Supervision and Skip-Connection

Training a DRCN is notably challenging due to exploding and vanishing gradients. To alleviate these issues, the authors introduce two key extensions:

  1. Recursive-Supervision: Every recursion is supervised, and intermediate high-resolution (HR) predictions are averaged to produce the final output. This helps mitigate gradient problems and improves convergence stability.
  2. Skip-Connection: Direct connections from the input to the reconstruction layers ensure a better retention of low-resolution (LR) information, which is crucial as LR and HR images share substantial information. This addition further enhances learning efficiency and prediction accuracy.

Experimental Results

The authors conducted extensive experiments to validate their methods, using datasets such as Set5, Set14, B100, and Urban100. The results show that their method outperforms state-of-the-art methods such as A+, SRCNN, RFL, and SelfEx across various scales (×2, ×3, ×4). The numerical results are highly commendable, with DRCN displaying superior PSNR and SSIM metrics across all tested benchmarks.

Notable Metrics

  • Set5 (×2 scale):
    • PSNR: DRCN - 37.63; SRCNN - 36.66
    • SSIM: DRCN - 0.9588; SRCNN - 0.9542
  • Urban100 (×4 scale):
    • PSNR: DRCN - 25.14; SelfEx - 24.79
    • SSIM: DRCN - 0.7510; SelfEx - 0.7374

The visual results also indicate that the DRCN produces sharper and more detailed images, particularly in challenging areas, where other methods blur the edges.

Implications and Future Directions

Practical Implications

The proposed DRCN addresses a critical trade-off in deep learning for image SR: the balance between model complexity and performance. By using recursive layers and recursive-supervision, it offers an efficient method for leveraging deep models without prohibitive computational costs. This method can be particularly impactful in applications requiring high-quality image scaling, such as medical imaging, remote sensing, and media enhancement.

Theoretical Implications

The recursive application of convolutional layers and the introduction of recursive-supervision and skip-connection provide new insights into handling deep networks' training dynamics. These contributions could be generalized to other domains within computer vision and beyond, particularly for tasks that require capturing extensive contextual information.

Future Directions

The paper suggests that further increasing the number of recursions might yield even better results, extending the depth to exploit image-level contexts fully. Moreover, the approach could be adapted for other image restoration tasks, such as denoising or compression artifact removal, suggesting a broad potential for future research and application.

Conclusion

This paper makes significant advances in the field of image super-resolution by introducing a deeply-recursive convolutional network with recursive-supervision and skip-connection. The proposed method achieves state-of-the-art results on several standard benchmarks, offering both theoretical innovations and practical improvements. As deep learning continues to evolve, the techniques introduced here will likely inspire further research and development in related areas of image processing and computer vision.

Whiteboard

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.