ARIN: Adaptive Resampling and Instance Normalization for Robust Blind Inpainting of Dunhuang Cave Paintings (2402.16188v1)
Abstract: Image enhancement algorithms are very useful for real world computer vision tasks where image resolution is often physically limited by the sensor size. While state-of-the-art deep neural networks show impressive results for image enhancement, they often struggle to enhance real-world images. In this work, we tackle a real-world setting: inpainting of images from Dunhuang caves. The Dunhuang dataset consists of murals, half of which suffer from corrosion and aging. These murals feature a range of rich content, such as Buddha statues, bodhisattvas, sponsors, architecture, dance, music, and decorative patterns designed by different artists spanning ten centuries, which makes manual restoration challenging. We modify two different existing methods (CAR, HINet) that are based upon state-of-the-art (SOTA) super resolution and deblurring networks. We show that those can successfully inpaint and enhance these deteriorated cave paintings. We further show that a novel combination of CAR and HINet, resulting in our proposed inpainting network (ARIN), is very robust to external noise, especially Gaussian noise. To this end, we present a quantitative and qualitative comparison of our proposed approach with existing SOTA networks and winners of the Dunhuang challenge. One of the proposed methods HINet) represents the new state of the art and outperforms the 1st place of the Dunhuang Challenge, while our combination ARIN, which is robust to noise, is comparable to the 1st place. We also present and discuss qualitative results showing the impact of our method for inpainting on Dunhuang cave images.
- F. Jinshi, “The caves of dunhuang,” Scala Arts & Heritage, 2010.
- “Mogao caves,” http://whc.unesco.org/en/list/440, last Accessed: 2022-01-06.
- X. Wang, N. Song, L. Zhang, and Y. Jiang, “Understanding subjects contained in dunhuang mural images for deep semantic annotation,” Journal of documentation, 2018.
- E. Panofsky, “Note on the importance of iconographical exactitude,” 1939.
- T. Yu, S. Zhang, C. Lin, S. You, J. Wu, J. Zhang, X. Ding, and H. An, “Dunhuang grottoes painting dataset and benchmark,” arXiv preprint arXiv:1907.04589, 2019.
- W. Sun and Z. Chen, “Learned image downscaling for upscaling using content adaptive resampler,” IEEE Transactions on Image Processing, vol. 29, pp. 4027–4040, 2020.
- H.-L. Wang, P.-H. Han, Y.-M. Chen, K.-W. Chen, X. Lin, M.-S. Lee, and Y.-P. Hung, “Dunhuang mural restoration using deep learning,” in SIGGRAPH Asia 2018 Technical Briefs, 2018, pp. 1–4.
- I.-M. Ciortan, S. George, and J. Y. Hardeberg, “Colour-balanced edge-guided digital inpainting: Applications on artworks,” Sensors, vol. 21, no. 6, 2021.
- J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei, “Imagenet: A large-scale hierarchical image database,” in 2009 IEEE conference on computer vision and pattern recognition. Ieee, 2009, pp. 248–255.
- T.-Y. Lin, M. Maire, S. Belongie, J. Hays, P. Perona, D. Ramanan, P. Dollár, and C. L. Zitnick, “Microsoft coco: Common objects in context,” in European conference on computer vision. Springer, 2014, pp. 740–755.
- L. Chen, X. Lu, J. Zhang, X. Chu, and C. Chen, “Hinet: Half instance normalization network for image restoration,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 182–192.
- O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in International Conference on Medical image computing and computer-assisted intervention. Springer, 2015, pp. 234–241.
- Z. Hui, X. Wang, and X. Gao, “Fast and accurate single image super-resolution via information distillation network,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 723–731.
- A. Ignatov, R. Timofte et al., “Pirm challenge on perceptual image enhancement on smartphones: report,” in European Conference on Computer Vision (ECCV) Workshops, January 2019.
- B. Lim, S. Son, H. Kim, S. Nah, and K. M. Lee, “Enhanced deep residual networks for single image super-resolution,” in The IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, July 2017.
- R. Zhang, P. Isola, A. A. Efros, E. Shechtman, and O. Wang, “The unreasonable effectiveness of deep features as a perceptual metric,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 586–595.
- Alexander Schmidt (36 papers)
- Prathmesh Madhu (13 papers)
- Andreas Maier (394 papers)
- Vincent Christlein (60 papers)
- Ronak Kosti (10 papers)