Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
162 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Multi-scale Progressive Feature Embedding for Accurate NIR-to-RGB Spectral Domain Translation (2312.16040v1)

Published 26 Dec 2023 in cs.CV and eess.IV

Abstract: NIR-to-RGB spectral domain translation is a challenging task due to the mapping ambiguities, and existing methods show limited learning capacities. To address these challenges, we propose to colorize NIR images via a multi-scale progressive feature embedding network (MPFNet), with the guidance of grayscale image colorization. Specifically, we first introduce a domain translation module that translates NIR source images into the grayscale target domain. By incorporating a progressive training strategy, the statistical and semantic knowledge from both task domains are efficiently aligned with a series of pixel- and feature-level consistency constraints. Besides, a multi-scale progressive feature embedding network is designed to improve learning capabilities. Experiments show that our MPFNet outperforms state-of-the-art counterparts by 2.55 dB in the NIR-to-RGB spectral domain translation task in terms of PSNR.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (29)
  1. F. Christnacher, E. Bacher, N. Metzger, S. Schertzer, Y. Lutz, J.-M. Poyet, and M. Laurenzis, “Portable bi-λ𝜆\lambdaitalic_λ SWIR/NIR GV gated viewing system for surveillance and security applications,” in Electro-Optical Remote Sensing XII, vol. 10796.   International Society for Optics and Photonics, 2018, pp. 54–64.
  2. Z. Liu, J. Wu, L. Fu, Y. Majeed, Y. Feng, R. Li, and Y. Cui, “Improved kiwifruit detection using pre-trained vgg16 with RGB and NIR information fusion,” IEEE Access, vol. 8, pp. 2327–2336, 2019.
  3. E. Protopapadakis, A. Doulamis, N. Doulamis, and E. Maltezos, “Stacked autoencoders driven by semi-supervised learning for building extraction from near infrared remote sensing imagery,” Remote Sensing, vol. 13, no. 3, p. 371, 2021.
  4. A. Deshpande, J. Rock, and D. Forsyth, “Learning large-scale automatic image colorization,” in IEEE International Conference on Computer Vision, 2015, pp. 567–575.
  5. Z. Cheng, Q. Yang, and B. Sheng, “Deep colorization,” in Proceedings of the IEEE international conference on computer vision, 2015, pp. 415–423.
  6. Y. Zhao, L.-M. Po, K.-W. Cheung, W.-Y. Yu, and Y. A. U. Rehman, “Scgan: Saliency map-guided colorization with generative adversarial network,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 31, no. 8, pp. 3062–3077, 2021.
  7. Z. Yang and Z. Chen, “Learning from paired and unpaired data: Alternately trained CycleGAN for near infrared image colorization,” in IEEE International Conference on Visual Communications and Image Processing, 2020, pp. 467–470.
  8. D. Valsesia, G. Fracastoro, and E. Magli, “NIR image colorization with graph-convolutional neural networks,” in IEEE International Conference on Visual Communications and Image Processing, 2020, pp. 451–454.
  9. L. Yan, X. Wang, M. Zhao, S. Liu, and J. Chen, “A multi-model fusion framework for NIR-to-RGB translation,” in IEEE International Conference on Visual Communications and Image Processing, 2020, pp. 459–462.
  10. P. L. Suárez, A. D. Sappa, and B. X. Vintimilla, “Learning to colorize infrared images,” in International Conference on Trends in Cyber-Physical Multi-Agent Systems, 2018, pp. 164–172.
  11. P. L. Suárez, A. D. Sappa, and B. X. Vintimilla, “Infrared image colorization based on a triplet DCGAN architecture,” in IEEE Conference on Computer Vision and Pattern Recognition Workshops, 2017, pp. 212–217.
  12. F. Wang, L. Liu, and C. Jung, “Deep near infrared colorization with semantic segmentation and transfer learning,” in IEEE International Conference on Visual Communications and Image Processing, 2020, pp. 455–458.
  13. W. Liang, D. Ding, and G. Wei, “An improved dualgan for near-infrared image colorization,” Infrared Physics & Technology, vol. 116, p. 103764, 2021.
  14. P. L. Suárez, A. D. Sappa, and B. X. Vintimilla, “Infrared image colorization based on a triplet dcgan architecture,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, 2017, pp. 18–23.
  15. Z. Dong, S.-i. Kamata, and T. P. Breckon, “Infrared image colorization using a s-shape network,” in 2018 25th IEEE International Conference on Image Processing.   IEEE, 2018, pp. 2242–2246.
  16. A. Mehri and A. D. Sappa, “Colorizing near infrared images through a cyclic adversarial approach of unpaired samples,” in IEEE Conference on Computer Vision and Pattern Recognition Workshops, 2019, pp. 1–9.
  17. T. Sun, C. Jung, Q. Fu, and Q. Han, “Nir to rgb domain translation using asymmetric cycle generative adversarial networks,” IEEE Access, vol. 7, pp. 112 459–112 469, 2019.
  18. J.-Y. Zhu, T. Park, P. Isola, and A. A. Efros, “Unpaired image-to-image translation using cycle-consistent adversarial networks,” in IEEE international conference on computer vision, 2017, pp. 2223–2232.
  19. A. Mehri, P. B. Ardakani, and A. D. Sappa, “Mprnet: Multi-path residual network for lightweight image super resolution,” in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2021, pp. 2704–2713.
  20. A. Mehri and A. D. Sappa, “Colorizing near infrared images through a cyclic adversarial approach of unpaired samples,” in IEEE Conference on Computer Vision and Pattern Recognition Workshops, 2019, pp. 971–979.
  21. X. Yang, J. Chen, Z. Yang, and Z. Chen, “Attention-guided nir image colorization via adaptive fusion of semantic and texture clues,” arXiv preprint arXiv:2107.09237, 2021.
  22. X. Huang and S. Belongie, “Arbitrary style transfer in real-time with adaptive instance normalization,” in IEEE International Conference on Computer Vision, 2017, pp. 1501–1510.
  23. Y. Jing, X. Liu, Y. Ding, X. Wang, E. Ding, M. Song, and S. Wen, “Dynamic instance normalization for arbitrary style transfer,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, no. 04, 2020, pp. 4369–4376.
  24. J. Hu, L. Shen, and G. Sun, “Squeeze-and-excitation networks,” in IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 7132–7141.
  25. S. Woo, J. Park, J.-Y. Lee, and I. S. Kweon, “CBAM: Convolutional block attention module,” in European Conference on Computer Vision, 2018, pp. 1–17.
  26. H. Zhao, O. Gallo, I. Frosio, and J. Kautz, “Loss functions for image restoration with neural networks,” IEEE Transactions on computational imaging, vol. 3, no. 1, pp. 47–57, 2016.
  27. Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli, “Image quality assessment: from error visibility to structural similarity,” IEEE Transactions on Image Processing, vol. 13, no. 4, pp. 600–612, 2004.
  28. R. Zhang, P. Isola, A. A. Efros, E. Shechtman, and O. Wang, “The unreasonable effectiveness of deep features as a perceptual metric,” in IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 586–595.
  29. M. Brown and S. Süsstrunk, “Multi-spectral sift for scene category recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition.   IEEE, 2011, pp. 177–184.
Citations (1)

Summary

We haven't generated a summary for this paper yet.