Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

AIGCOIQA2024: Perceptual Quality Assessment of AI Generated Omnidirectional Images (2404.01024v1)

Published 1 Apr 2024 in cs.CV and eess.IV

Abstract: In recent years, the rapid advancement of Artificial Intelligence Generated Content (AIGC) has attracted widespread attention. Among the AIGC, AI generated omnidirectional images hold significant potential for Virtual Reality (VR) and Augmented Reality (AR) applications, hence omnidirectional AIGC techniques have also been widely studied. AI-generated omnidirectional images exhibit unique distortions compared to natural omnidirectional images, however, there is no dedicated Image Quality Assessment (IQA) criteria for assessing them. This study addresses this gap by establishing a large-scale AI generated omnidirectional image IQA database named AIGCOIQA2024 and constructing a comprehensive benchmark. We first generate 300 omnidirectional images based on 5 AIGC models utilizing 25 text prompts. A subjective IQA experiment is conducted subsequently to assess human visual preferences from three perspectives including quality, comfortability, and correspondence. Finally, we conduct a benchmark experiment to evaluate the performance of state-of-the-art IQA models on our database. The database will be released to facilitate future research.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (39)
  1. “Generative adversarial nets,” in Proceedings of the Advances in Neural Information Processing Systems, 2014, vol. 27.
  2. “Auto-Encoding Variational Bayes,” arXiv e-prints, p. arXiv:1312.6114, Dec. 2013.
  3. “High-resolution image synthesis with latent diffusion models,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 10684–10695.
  4. “BLIP-2: Bootstrapping Language-Image Pre-training with Frozen Image Encoders and Large Language Models,” arXiv e-prints, p. arXiv:2301.12597, Jan. 2023.
  5. “Learning transferable visual models from natural language supervision,” in Proceedings of the International Conference on Machine Learning. pp. 8748–8763, PMLR.
  6. “Hierarchical Text-Conditional Image Generation with CLIP Latents,” arXiv e-prints, p. arXiv:2204.06125, Apr. 2022.
  7. “High-resolution image synthesis with latent diffusion models,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2022, pp. 10684–10695.
  8. “MVDiffusion: Enabling holistic multi-view image generation with correspondence-aware diffusion,” in Proceedings of the Neural Information Processing Systems (NeurIPS), 2023.
  9. “Text2light: Zero-shot text-driven hdr panorama generation,” ACM Trans. Graph., vol. 41, no. 6, nov 2022.
  10. “GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium,” in Proceedings of the Advances in Neural Information Processing Systems, 2017, vol. 30.
  11. “CLIPScore: A Reference-free Evaluation Metric for Image Captioning,” arXiv e-prints, p. arXiv:2104.08718, Apr. 2021.
  12. “Aigciqa2023: A large-scale image quality assessment database for ai generated images: from the perspectives of quality, authenticity and correspondence,” in Proceedings of the CAAI International Conference on Artificial Intelligence (CICAI). Springer, 2023, pp. 46–57.
  13. “Agiqa-3k: An open database for ai-generated image quality assessment,” arXiv preprint arXiv:2306.04717, 2023.
  14. “Matterport3d: Learning from rgb-d data in indoor environments,” Proceedngs of the International Conference on 3D Vision (3DV), 2017.
  15. “Recognizing scene viewpoint using panoramic place representation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2012, pp. 2695–2702.
  16. “SceneScape: Text-Driven Consistent Scene Generation,” arXiv e-prints, p. arXiv:2302.01133, Feb. 2023.
  17. “Audio-visual Saliency for Omnidirectional Videos,” arXiv e-prints, p. arXiv:2311.05190, Nov. 2023.
  18. “Perceptual Quality Assessment of Omnidirectional Audio-Visual Signals,” in Proceedings of the CAAI International Conference on Artificial Intelligence (CICAI), Singapore, 2024, pp. 512–525.
  19. “Attentive deep image quality assessment for omnidirectional stitching,” IEEE Journal of Selected Topics in Signal Processing, vol. 17, no. 6, pp. 1150–1164, 2023.
  20. “Saliency in augmented reality,” in Proceedings of the 30th ACM International Conference on Multimedia, 2022.
  21. “Perceptual quality assessment of omnidirectional images,” in Proceedings of the IEEE International Symposium on Circuits and Systems (ISCAS), 2018, pp. 1–5.
  22. “Augmented reality image quality assessment based on visual confusion theory,” in Proceedings of the IEEE International Symposium on Broadband Multimedia Systems and Broadcasting (BMSB), 2022, pp. 1–6.
  23. “Viewing behavior supported visual saliency predictor for 360 degree videos,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 32, no. 7, pp. 4188–4201, 2022.
  24. “Ivqad 2017: An immersive video quality assessment database,” in Proceedings of the International Conference on Systems, Signals and Image Processing (IWSSIP), 2017, pp. 1–5.
  25. “Confusing image quality assessment: Toward better augmented reality experience,” IEEE Transactions on Image Processing, vol. 31, pp. 7206–7221, 2022.
  26. B. Series, “Methodology for the subjective assessment of the quality of television pictures,” Tech. Rep. ITU-R BT, ITU-R, 2012.
  27. “Learning without human scores for blind image quality assessment,” Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition.
  28. “Blind image quality estimation via distortion aggravation,” IEEE Transactions on Broadcasting, vol. 64, no. 2, pp. 508–517, 2018.
  29. “Making a “completely blind” image quality analyzer,” IEEE Signal Processing Letters, vol. 20, no. 3, pp. 209–212, 2013.
  30. “A feature-enriched completely blind image quality evaluator,” IEEE Transactions on Image Processing, vol. 24, no. 8, pp. 2579–2591, 2015.
  31. “Blind image quality assessment based on high order statistics aggregation,” IEEE Transactions on Image Processing, vol. 25, pp. 4444–4457, 2016.
  32. “Blind quality assessment based on pseudo-reference image,” IEEE Transactions on Multimedia, vol. 20, no. 8, pp. 2049–2062, 2018.
  33. “Fisblim: A five-step blind metric for quality assessment of multiply distorted images,” in Proceedings of the SiPS, 2013, pp. 241–246.
  34. “Convolutional neural networks for no-reference image quality assessment,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2014.
  35. “Deep residual learning for image recognition,” in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 770–778.
  36. “Very Deep Convolutional Networks for Large-Scale Image Recognition,” arXiv e-prints, p. arXiv:1409.1556, Sept. 2014.
  37. “Blindly assess image quality in the wild guided by a self-adaptive hyper network,” in 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020, pp. 3664–3673.
  38. “MANIQA: Multi-dimension Attention Network for No-Reference Image Quality Assessment,” arXiv e-prints, p. arXiv:2204.08958, Apr. 2022.
  39. “No-reference image quality assessment via transformers, relative ranking, and self-consistency,” in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), January 2022, pp. 1220–1230.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Liu Yang (195 papers)
  2. Huiyu Duan (38 papers)
  3. Long Teng (16 papers)
  4. Yucheng Zhu (20 papers)
  5. Xiaohong Liu (117 papers)
  6. Menghan Hu (21 papers)
  7. Xiongkuo Min (139 papers)
  8. Guangtao Zhai (231 papers)
  9. Patrick Le Callet (41 papers)
Citations (7)

Summary

We haven't generated a summary for this paper yet.