Papers
Topics
Authors
Recent
2000 character limit reached

Q-Refine: A Perceptual Quality Refiner for AI-Generated Image (2401.01117v1)

Published 2 Jan 2024 in cs.CV and eess.IV

Abstract: With the rapid evolution of the Text-to-Image (T2I) model in recent years, their unsatisfactory generation result has become a challenge. However, uniformly refining AI-Generated Images (AIGIs) of different qualities not only limited optimization capabilities for low-quality AIGIs but also brought negative optimization to high-quality AIGIs. To address this issue, a quality-award refiner named Q-Refine is proposed. Based on the preference of the Human Visual System (HVS), Q-Refine uses the Image Quality Assessment (IQA) metric to guide the refining process for the first time, and modify images of different qualities through three adaptive pipelines. Experimental shows that for mainstream T2I models, Q-Refine can perform effective optimization to AIGIs of different qualities. It can be a general refiner to optimize AIGIs from both fidelity and aesthetic quality levels, thus expanding the application of the T2I generation models.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (32)
  1. “Agiqa-3k: An open database for ai-generated image quality assessment,” IEEE TCSVT, 2023.
  2. “Adversarial text-to-image synthesis: A review,” Neural Networks, 2021.
  3. “Text-to-image diffusion model in generative ai: A survey,” arXiv:2303.07909, 2023.
  4. “Unsupervised real-world image super resolution via domain-distance aware training,” in IEEE/CVF CVPR, 2021.
  5. “Diffbir: Towards blind image restoration with generative diffusion prior,” arXiv:2308.15070, 2023.
  6. “Residual feature distillation network for lightweight image super-resolution,” in ECCV, 2020.
  7. “High-resolution image synthesis with latent diffusion models,” in IEEE CVPR, 2022.
  8. “Text-guided synthesis of artistic images with retrieval-augmented diffusion models,” arXiv:2207.13038, 2022.
  9. “Stablevqa: A deep no-reference quality assessment model for video stability,” in ACM MM, 2023.
  10. “Vdpve: Vqa dataset for perceptual video enhancement,” in IEEE/CVF CVPR, 2023.
  11. “A real-time blind quality-of-experience assessment metric for http adaptive streaming,” in IEEE ICME, 2023.
  12. “A full-reference quality assessment metric for cartoon images,” in IEEE MMSP, 2022.
  13. “Xgc-vqa: A unified video quality assessment model for user, professionally, and occupationally-generated content,” in IEEE ICMEW, 2023.
  14. “Q-bench: A benchmark for general-purpose foundation models on low-level vision,” arXiv:2309.14181, 2023.
  15. “Q-instruct: Improving low-level visual abilities for multi-modality foundation models,” arXiv:2311.06783, 2023.
  16. “Q-boost: On visual quality assessment ability of low-level multi-modality foundation models,” arXiv:2312.15300, 2023.
  17. “Advancing zero-shot digital human quality assessment through text-prompted evaluation,” arXiv:2307.02808, 2023.
  18. “Gms-3dqa: Projection-based grid mini-patch sampling for 3d model quality assessment,” arXiv:2306.05658, 2023.
  19. “From patches to pictures (paq-2-piq): Mapping the perceptual space of picture quality,” in IEEE/CVF CVPR, 2020.
  20. “Faster r-cnn: Towards real-time object detection with region proposal networks,” NIPS, 2015.
  21. “Pixart-α𝛼\alphaitalic_α: Fast training of diffusion transformer for photorealistic text-to-image synthesis,” arXiv:2310.00426, 2023.
  22. Dianyuan Han, “Comparison of commonly used image interpolation methods,” in ICCSEE, 2013.
  23. “Best prompts for text-to-image models and how to find them,” in ACM SIGIRl, 2023.
  24. “Glide: Towards photorealistic image generation and editing with text-guided diffusion models,” in ICML, 2022.
  25. “Hierarchical text-conditional image generation with clip latents,” arXiv:2204.06125, 2022.
  26. David Holz, “Midjourney,” https://www.midjourney.com/, 2023.
  27. “A perceptual quality assessment exploration for aigc images,” in IEEE ICMEW, 2023.
  28. “Aigciqa2023: A large-scale image quality assessment database for ai generated images: from the perspectives of quality, authenticity and correspondence,” in CICAI, 2023.
  29. “Gans trained by a two time-scale update rule converge to a local nash equilibrium,” in NIPS, 2017.
  30. “No-reference image quality assessment in the spatial domain,” IEEE TIP, 2012.
  31. “Blindly assess image quality in the wild guided by a self-adaptive hyper network,” in IEEE/CVF CVPR, 2020.
  32. “Exploring clip for assessing the look and feel of images,” in AAAI, 2023.
Citations (12)

Summary

We haven't generated a summary for this paper yet.

Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 1 tweet and received 0 likes.

Upgrade to Pro to view all of the tweets about this paper: