Papers
Topics
Authors
Recent
2000 character limit reached

Demystifying the Visual Quality Paradox in Multimodal Large Language Models (2506.15645v1)

Published 18 Jun 2025 in cs.CV and cs.AI

Abstract: Recent Multimodal LLMs (MLLMs) excel on benchmark vision-language tasks, yet little is known about how input visual quality shapes their responses. Does higher perceptual quality of images already translate to better MLLM understanding? We conduct the first systematic study spanning leading MLLMs and a suite of vision-language benchmarks, applying controlled degradations and stylistic shifts to each image. Surprisingly, we uncover a visual-quality paradox: model, task, and even individual-instance performance can improve when images deviate from human-perceived fidelity. Off-the-shelf restoration pipelines fail to reconcile these idiosyncratic preferences. To close the gap, we introduce Visual-Quality Test-Time Tuning (VQ-TTT)-a lightweight adaptation module that: (1) inserts a learnable, low-rank kernel before the frozen vision encoder to modulate frequency content; and (2) fine-tunes only shallow vision-encoder layers via LoRA. VQ-TTT dynamically adjusts each input image in a single forward pass, aligning it with task-specific model preferences. Across the evaluated MLLMs and all datasets, VQ-TTT lifts significant average accuracy, with no external models, cached features, or extra training data. These findings redefine better'' visual inputs for MLLMs and highlight the need for adaptive, rather than universallyclean'', imagery, in the new era of AI being the main data customer.

Summary

Demystifying the Visual Quality Paradox in Multimodal LLMs

Overview

The paper "Demystifying the Visual Quality Paradox in Multimodal LLMs" embarks on a detailed exploration of how image quality impacts the performance of Multimodal LLMs (MLLMs). The authors conduct a systematic study across leading MLLMs and various vision-language benchmarks, examining how different image alterations affect task performance. Surprisingly, they identify a paradox where decreased visual fidelity can result in enhanced MLLM capabilities, challenging the prevalent assumption that higher image quality invariably improves model understanding.

Key Insights and Methodology

The research highlights a "visual-quality paradox," where MLLMs may perform better with images that are perceived by humans as being of lower quality. The investigation involved applying controlled degradations and stylistic shifts to images and assessing the performance across a range of models and vision-language tasks. Despite expectations, certain tasks and models showed improved performance with degraded input images. The authors attribute this paradox to a misalignment between human-centered image quality metrics and those that benefit models.

To address these idiosyncratic behaviors in MLLMs, the paper introduces the Visual-Quality Test-Time Tuning (VQ-TTT) framework. This novel adaptation module dynamically optimizes the input visual quality without altering the MLLM's foundational structure. VQ-TTT uses a learnable, low-rank kernel placed ahead of the frozen vision encoder to adjust the frequency content of the input images, complemented by LoRA-based fine-tuning on shallow layers of the vision encoder. This approach tailors input images to meet specific task and model requirements, offering consistent performance improvements across numerous datasets and MLLM architectures.

Experimental Results

Empirical analysis conducted in the study involved various types of image degradation, including noise, motion blur, defocus blur, snow, and fog, applied at multiple intensities across 13 distinct datasets. Notably, some models, such as LLaVA and Qwen, displayed non-trivial performance gains on reasoning and comprehension tasks with visually degraded images. This suggests certain visual perturbations may enhance task-specific features that are vital for the models' internal computations.

The application of high-quality image restoration techniques, including advanced pipelined and co-trained restoration models, often failed to bridge the performance gap or even negatively impacted accuracy. This underlines the necessity for model-aligned, task-specific visual adaptation.

Implications and Future Directions

The findings of this paper suggest a significant shift in understanding what constitutes "better" visual inputs for MLLMs. It emphasizes the need for input-adaptive methodologies rather than universally pristine visual imagery. VQ-TTT exemplifies a lightweight strategy capable of aligning visual inputs with the nuanced preferences of MLLMs across different tasks and contexts.

Practically, this research encourages the development of more adaptable MLLMs that can effectively process diverse real-world image inputs. Theoretically, it challenges current paradigms and compels a reconsideration of fidelity metrics tailored specifically to machine vision. Future developments in AI may involve evolving adaptive models that integrate intrinsic visual alignment during both the training and deployment phases, ultimately enhancing the robustness and versatility of vision-language systems.

In conclusion, the work offers a refined perspective on input quality for MLLMs, emphasizing adaptability and alignment over conventional quality metrics, thereby fostering broader practical applications and enhancements in AI-driven visual comprehension.

Whiteboard

Video Overview

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.