Perception- and Fidelity-aware Reduced-Reference Super-Resolution Image Quality Assessment (2405.09472v2)
Abstract: With the advent of image super-resolution (SR) algorithms, how to evaluate the quality of generated SR images has become an urgent task. Although full-reference methods perform well in SR image quality assessment (SR-IQA), their reliance on high-resolution (HR) images limits their practical applicability. Leveraging available reconstruction information as much as possible for SR-IQA, such as low-resolution (LR) images and the scale factors, is a promising way to enhance assessment performance for SR-IQA without HR for reference. In this letter, we attempt to evaluate the perceptual quality and reconstruction fidelity of SR images considering LR images and scale factors. Specifically, we propose a novel dual-branch reduced-reference SR-IQA network, \ie, Perception- and Fidelity-aware SR-IQA (PFIQA). The perception-aware branch evaluates the perceptual quality of SR images by leveraging the merits of global modeling of Vision Transformer (ViT) and local relation of ResNet, and incorporating the scale factor to enable comprehensive visual perception. Meanwhile, the fidelity-aware branch assesses the reconstruction fidelity between LR and SR images through their visual perception. The combination of the two branches substantially aligns with the human visual system, enabling a comprehensive SR image evaluation. Experimental results indicate that our PFIQA outperforms current state-of-the-art models across three widely-used SR-IQA benchmarks. Notably, PFIQA excels in assessing the quality of real-world SR images.
- H. Wu, J. Chen, T. Wang, X. Lai, and J. Cao, “Ship license plate super-resolution in the wild,” IEEE Signal Processing Letters, vol. 30, pp. 394–398, 2023.
- M.-I. Georgescu, R. T. Ionescu, A.-I. Miron, O. Savencu, N.-C. Ristea, N. Verga, and F. S. Khan, “Multimodal multi-head convolutional attention with various kernel sizes for medical image super-resolution,” in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2023, pp. 2195–2205.
- Y. Xiao, Q. Yuan, K. Jiang, J. He, X. Jin, and L. Zhang, “EDiffSR: An efficient diffusion probabilistic model for remote sensing image super-resolution,” IEEE Transactions on Geoscience and Remote Sensing, vol. 62, pp. 1–14, 2024.
- J. Shin, Y.-H. Jo, B.-K. Khim, and S. M. Kim, “U-net super-resolution model of goci to goci-ii image conversion,” IEEE Transactions on Geoscience and Remote Sensing, vol. 62, pp. 1–12, 2024.
- R. Neshatavar, M. Yavartanoo, S. Son, and K. M. Lee, “ICF-SRSR: Invertible scale-conditional function for self-supervised real-world single image super-resolution,” in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2024, pp. 1557–1567.
- H. Chen, L. Dong, H. Yang, X. He, and C. Zhu, “Unsupervised real-world image super-resolution via dual synthetic-to-realistic and realistic-to-synthetic translations,” IEEE Signal Processing Letters, vol. 29, pp. 1282–1286, 2022.
- J. Wu, Y. Wang, and X. Zhang, “Lightweight asymmetric convolutional distillation network for single image super-resolution,” IEEE Signal Processing Letters, vol. 30, pp. 733–737, 2023.
- Y. Wang and T. Zhang, “OSFFNet: Omni-stage feature fusion network for lightweight image super-resolution,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 38, no. 6, 2024, pp. 5660–5668.
- Y. Zhao, Q. Teng, H. Chen, S. Zhang, X. He, Y. Li, and R. E. Sheriff, “Activating more information in arbitrary-scale image super-resolution,” IEEE Transactions on Multimedia, vol. 26, pp. 7946–7961, 2024.
- Y. Zhou, L. Gao, Z. Tang, and B. Wei, “Recognition-guided diffusion model for scene text image super-resolution,” in IEEE International Conference on Acoustics, Speech and Signal Processing, 2024, pp. 2940–2944.
- C. Noguchi, S. Fukuda, and M. Yamanaka, “Scene text image super-resolution based on text-conditional diffusion models,” in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2024, pp. 1485–1495.
- A. Mittal, A. K. Moorthy, and A. C. Bovik, “No-reference image quality assessment in the spatial domain,” IEEE Transactions on Image Processing, vol. 21, no. 12, pp. 4695–4708, 2012.
- A. Mittal, R. Soundararajan, and A. C. Bovik, “Making a “completely blind” image quality analyzer,” IEEE Signal Processing Letters, vol. 20, no. 3, pp. 209–212, 2012.
- S. Yang, T. Wu, S. Shi, S. Lao, Y. Gong, M. Cao, J. Wang, and Y. Yang, “MANIQA: Multi-dimension attention network for no-reference image quality assessment,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2022, pp. 1190–1199.
- M. Cheon, S.-J. Yoon, B. Kang, and J. Lee, “Perceptual image quality assessment with transformers,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 433–442.
- S. Lao, Y. Gong, S. Shi, S. Yang, T. Wu, J. Wang, W. Xia, and Y. Yang, “Attentions help cnns see better: Attention-based hybrid image quality assessment network,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 1140–1149.
- L. Yu, J. Li, F. Pakdaman, M. Ling, and M. Gabbouj, “MAMIQA: No-reference image quality assessment based on multiscale attention mechanism with natural scene statistics,” IEEE Signal Processing Letters, vol. 30, pp. 588–592, 2023.
- K. Zhang, T. Zhao, W. Chen, Y. Niu, J. Hu, and W. Lin, “Perception-driven similarity-clarity tradeoff for image super-resolution quality assessment,” IEEE Transactions on Circuits and Systems for Video Technology, 2023.
- T. Zhao, Y. Lin, Y. Xu, W. Chen, and Z. Wang, “Learning-based quality assessment for image super-resolution,” IEEE Transactions on Multimedia, vol. 24, pp. 3570–3581, 2021.
- J. Fu, “Scale guided hypernetwork for blind super-resolution image quality assessment,” arXiv preprint arXiv:2306.02398, 2023.
- H. Li, K. Zhang, Z. Niu, and H. Shi, “C2MT: A credible and class-aware multi-task transformer for sr-iqa,” IEEE Signal Processing Letters, vol. 29, pp. 2662–2666, 2022.
- K. Zhang, T. Zhao, W. Chen, Y. Niu, and J. Hu, “SPQE: Structure-and-perception-based quality evaluation for image super-resolution,” arXiv preprint arXiv:2205.03584, 2022.
- X. Huang, W. Li, J. Hu, H. Chen, and Y. Wang, “RefSR-NeRF: Towards high fidelity and super resolution view synthesis,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 8244–8253.
- X. Luo, Y. Xie, Y. Qu, and Y. Fu, “SkipDiff: Adaptive skip diffusion model for high-fidelity perceptual image super-resolution,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 38, no. 5, 2024, pp. 4017–4025.
- F. Zhou, W. Sheng, Z. Lu, B. Kang, M. Chen, and G. Qiu, “Super-resolution image visual quality assessment based on structure–texture features,” Signal Processing: Image Communication, vol. 117, p. 117025, 2023.
- A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly et al., “An image is worth 16x16 words: Transformers for image recognition at scale,” arXiv preprint arXiv:2010.11929, 2020.
- K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 770–778.
- A. Saha, S. Mishra, and A. C. Bovik, “Re-IQA: Unsupervised learning for image quality assessment in the wild,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 5846–5855.
- C. Chen, J. Mo, J. Hou, H. Wu, L. Liao, W. Sun, Q. Yan, and W. Lin, “TOPIQ: A top-down approach from semantics to distortions for image quality assessment,” IEEE Transactions on Image Processing, 2024.
- F. Zhou, R. Yao, B. Liu, and G. Qiu, “Visual quality assessment for super-resolved images: Database and method,” IEEE Transactions on Image Processing, vol. 28, no. 7, pp. 3528–3541, 2019.
- H. Yeganeh, M. Rostami, and Z. Wang, “Objective quality assessment of interpolated natural images,” IEEE Transactions on Image Processing, vol. 24, no. 11, pp. 4651–4663, 2015.
- Q. Jiang, Z. Liu, K. Gu, F. Shao, X. Zhang, H. Liu, and W. Lin, “Single image super-resolution quality assessment: a real-world dataset, subjective studies, and an objective metric,” IEEE Transactions on Image Processing, vol. 31, pp. 2279–2294, 2022.
- L. Kang, P. Ye, Y. Li, and D. Doermann, “Convolutional neural networks for no-reference image quality assessment,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2014, pp. 1733–1740.
- C. Ma, C.-Y. Yang, X. Yang, and M.-H. Yang, “Learning a no-reference quality metric for single-image super-resolution,” Computer Vision and Image Understanding, vol. 158, pp. 1–16, 2017.
- W. Zhou and Z. Wang, “Quality assessment of image super-resolution: Balancing deterministic and statistical fidelity,” in Proceedings of the 30th ACM International Conference on Multimedia, 2022, pp. 934–942.
- O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy, A. Khosla, M. Bernstein, A. C. Berg, and L. Fei-Fei, “Imagenet large scale visual recognition challenge,” International Journal of Computer Vision, vol. 115, pp. 211–252, 2015.
- Xinying Lin (1 paper)
- Xuyang Liu (23 papers)
- Hong Yang (78 papers)
- Xiaohai He (14 papers)
- Honggang Chen (21 papers)