Multi-scale Progressive Feature Embedding for Accurate NIR-to-RGB Spectral Domain Translation (2312.16040v1)
Abstract: NIR-to-RGB spectral domain translation is a challenging task due to the mapping ambiguities, and existing methods show limited learning capacities. To address these challenges, we propose to colorize NIR images via a multi-scale progressive feature embedding network (MPFNet), with the guidance of grayscale image colorization. Specifically, we first introduce a domain translation module that translates NIR source images into the grayscale target domain. By incorporating a progressive training strategy, the statistical and semantic knowledge from both task domains are efficiently aligned with a series of pixel- and feature-level consistency constraints. Besides, a multi-scale progressive feature embedding network is designed to improve learning capabilities. Experiments show that our MPFNet outperforms state-of-the-art counterparts by 2.55 dB in the NIR-to-RGB spectral domain translation task in terms of PSNR.
- F. Christnacher, E. Bacher, N. Metzger, S. Schertzer, Y. Lutz, J.-M. Poyet, and M. Laurenzis, “Portable bi-λ𝜆\lambdaitalic_λ SWIR/NIR GV gated viewing system for surveillance and security applications,” in Electro-Optical Remote Sensing XII, vol. 10796. International Society for Optics and Photonics, 2018, pp. 54–64.
- Z. Liu, J. Wu, L. Fu, Y. Majeed, Y. Feng, R. Li, and Y. Cui, “Improved kiwifruit detection using pre-trained vgg16 with RGB and NIR information fusion,” IEEE Access, vol. 8, pp. 2327–2336, 2019.
- E. Protopapadakis, A. Doulamis, N. Doulamis, and E. Maltezos, “Stacked autoencoders driven by semi-supervised learning for building extraction from near infrared remote sensing imagery,” Remote Sensing, vol. 13, no. 3, p. 371, 2021.
- A. Deshpande, J. Rock, and D. Forsyth, “Learning large-scale automatic image colorization,” in IEEE International Conference on Computer Vision, 2015, pp. 567–575.
- Z. Cheng, Q. Yang, and B. Sheng, “Deep colorization,” in Proceedings of the IEEE international conference on computer vision, 2015, pp. 415–423.
- Y. Zhao, L.-M. Po, K.-W. Cheung, W.-Y. Yu, and Y. A. U. Rehman, “Scgan: Saliency map-guided colorization with generative adversarial network,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 31, no. 8, pp. 3062–3077, 2021.
- Z. Yang and Z. Chen, “Learning from paired and unpaired data: Alternately trained CycleGAN for near infrared image colorization,” in IEEE International Conference on Visual Communications and Image Processing, 2020, pp. 467–470.
- D. Valsesia, G. Fracastoro, and E. Magli, “NIR image colorization with graph-convolutional neural networks,” in IEEE International Conference on Visual Communications and Image Processing, 2020, pp. 451–454.
- L. Yan, X. Wang, M. Zhao, S. Liu, and J. Chen, “A multi-model fusion framework for NIR-to-RGB translation,” in IEEE International Conference on Visual Communications and Image Processing, 2020, pp. 459–462.
- P. L. Suárez, A. D. Sappa, and B. X. Vintimilla, “Learning to colorize infrared images,” in International Conference on Trends in Cyber-Physical Multi-Agent Systems, 2018, pp. 164–172.
- P. L. Suárez, A. D. Sappa, and B. X. Vintimilla, “Infrared image colorization based on a triplet DCGAN architecture,” in IEEE Conference on Computer Vision and Pattern Recognition Workshops, 2017, pp. 212–217.
- F. Wang, L. Liu, and C. Jung, “Deep near infrared colorization with semantic segmentation and transfer learning,” in IEEE International Conference on Visual Communications and Image Processing, 2020, pp. 455–458.
- W. Liang, D. Ding, and G. Wei, “An improved dualgan for near-infrared image colorization,” Infrared Physics & Technology, vol. 116, p. 103764, 2021.
- P. L. Suárez, A. D. Sappa, and B. X. Vintimilla, “Infrared image colorization based on a triplet dcgan architecture,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, 2017, pp. 18–23.
- Z. Dong, S.-i. Kamata, and T. P. Breckon, “Infrared image colorization using a s-shape network,” in 2018 25th IEEE International Conference on Image Processing. IEEE, 2018, pp. 2242–2246.
- A. Mehri and A. D. Sappa, “Colorizing near infrared images through a cyclic adversarial approach of unpaired samples,” in IEEE Conference on Computer Vision and Pattern Recognition Workshops, 2019, pp. 1–9.
- T. Sun, C. Jung, Q. Fu, and Q. Han, “Nir to rgb domain translation using asymmetric cycle generative adversarial networks,” IEEE Access, vol. 7, pp. 112 459–112 469, 2019.
- J.-Y. Zhu, T. Park, P. Isola, and A. A. Efros, “Unpaired image-to-image translation using cycle-consistent adversarial networks,” in IEEE international conference on computer vision, 2017, pp. 2223–2232.
- A. Mehri, P. B. Ardakani, and A. D. Sappa, “Mprnet: Multi-path residual network for lightweight image super resolution,” in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2021, pp. 2704–2713.
- A. Mehri and A. D. Sappa, “Colorizing near infrared images through a cyclic adversarial approach of unpaired samples,” in IEEE Conference on Computer Vision and Pattern Recognition Workshops, 2019, pp. 971–979.
- X. Yang, J. Chen, Z. Yang, and Z. Chen, “Attention-guided nir image colorization via adaptive fusion of semantic and texture clues,” arXiv preprint arXiv:2107.09237, 2021.
- X. Huang and S. Belongie, “Arbitrary style transfer in real-time with adaptive instance normalization,” in IEEE International Conference on Computer Vision, 2017, pp. 1501–1510.
- Y. Jing, X. Liu, Y. Ding, X. Wang, E. Ding, M. Song, and S. Wen, “Dynamic instance normalization for arbitrary style transfer,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, no. 04, 2020, pp. 4369–4376.
- J. Hu, L. Shen, and G. Sun, “Squeeze-and-excitation networks,” in IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 7132–7141.
- S. Woo, J. Park, J.-Y. Lee, and I. S. Kweon, “CBAM: Convolutional block attention module,” in European Conference on Computer Vision, 2018, pp. 1–17.
- H. Zhao, O. Gallo, I. Frosio, and J. Kautz, “Loss functions for image restoration with neural networks,” IEEE Transactions on computational imaging, vol. 3, no. 1, pp. 47–57, 2016.
- Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli, “Image quality assessment: from error visibility to structural similarity,” IEEE Transactions on Image Processing, vol. 13, no. 4, pp. 600–612, 2004.
- R. Zhang, P. Isola, A. A. Efros, E. Shechtman, and O. Wang, “The unreasonable effectiveness of deep features as a perceptual metric,” in IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 586–595.
- M. Brown and S. Süsstrunk, “Multi-spectral sift for scene category recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition. IEEE, 2011, pp. 177–184.