Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

RankIQA: Learning from Rankings for No-reference Image Quality Assessment (1707.08347v1)

Published 26 Jul 2017 in cs.CV

Abstract: We propose a no-reference image quality assessment (NR-IQA) approach that learns from rankings (RankIQA). To address the problem of limited IQA dataset size, we train a Siamese Network to rank images in terms of image quality by using synthetically generated distortions for which relative image quality is known. These ranked image sets can be automatically generated without laborious human labeling. We then use fine-tuning to transfer the knowledge represented in the trained Siamese Network to a traditional CNN that estimates absolute image quality from single images. We demonstrate how our approach can be made significantly more efficient than traditional Siamese Networks by forward propagating a batch of images through a single network and backpropagating gradients derived from all pairs of images in the batch. Experiments on the TID2013 benchmark show that we improve the state-of-the-art by over 5%. Furthermore, on the LIVE benchmark we show that our approach is superior to existing NR-IQA techniques and that we even outperform the state-of-the-art in full-reference IQA (FR-IQA) methods without having to resort to high-quality reference images to infer IQA.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Xialei Liu (35 papers)
  2. Joost van de Weijer (133 papers)
  3. Andrew D. Bagdanov (47 papers)
Citations (395)

Summary

Insights on "RankIQA: Learning from Rankings for No-reference Image Quality Assessment"

The paper "RankIQA: Learning from Rankings for No-reference Image Quality Assessment" introduces a novel approach for tackling the No-reference Image Quality Assessment (NR-IQA) problem by leveraging the learning from image rankings, termed as RankIQA. The authors' primary objective is to address the challenge posed by the limited size of existing IQA datasets, which hampers the training of deep convolutional neural networks (CNNs). The methodology hinges on synthesizing rankings from distorted images and capitalizing on a Siamese Network to learn these rankings, followed by transferring this learned knowledge to a CNN capable of directly predicting image quality from single images.

The key innovation of this research is its method of generating training data. By introducing synthetic distortions of varying intensities onto images, the authors generate datasets where intra-set ranking is intrinsically known, removing the need for labor-intensive manual labeling. The process involves training a Siamese Network on these synthetic rankings, which is later fine-tuned on smaller, real-world IQA datasets. This two-step training strategy enables the resultant model to generalize well for NR-IQA tasks even in the absence of large labeled datasets typically required by deep networks.

A notable contribution is the improved efficiency in training Siamese Networks. Traditional Siamese Networks demand recalculating gradients for each image pair, leading to redundant computations. The authors optimize this process by passing a batch of images through the network's forward pass and performing backpropagation using gradients derived from all possible image pairs within the batch. This innovation significantly reduces computation times, making the training process more practical without compromising on performance.

From an empirical standpoint, the performance improvements achieved using RankIQA are impressive. The experiments on TID2013 and LIVE benchmarks demonstrate a performance leap over existing NR-IQA methods. On the TID2013 dataset, RankIQA outperformed the state-of-the-art by approximately 5%, and in some instances, surpassed full-reference IQA (FR-IQA) methods on the LIVE dataset. An important aspect of their results is the cross-dataset generality, evidenced by satisfactory results even when training ranking networks on an unrelated dataset like Places2, highlighting the robustness of the learned representations.

The implications of this work are manifold. Practically, it offers a scalable solution to NR-IQA, reducing the reliance on extensive labeled datasets and enabling the efficient training of deep architectures. Theoretically, it presents a refined approach to using synthetically generated data for model training, potentially applicable to other domains facing similar data scarcity challenges. Moreover, by establishing that NR-IQA can compete with and exceed FR-IQA methods, the research underscores the feasibility of no-reference methods in real-world applications where reference images are unavailable.

Future developments could explore extending the approach to more complex distortion categories or further refining the efficiency of the training process. Additionally, as deep learning architectures continue to evolve, integrating RankIQA with next-generation models could present new opportunities for pushing the boundaries of image quality assessment. Exploring joint training with other computer vision tasks could also uncover synergistic benefits, enhancing overall image analysis capabilities.