Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Where do Large Vision-Language Models Look at when Answering Questions? (2503.13891v1)

Published 18 Mar 2025 in cs.CV and cs.CL

Abstract: Large Vision-LLMs (LVLMs) have shown promising performance in vision-language understanding and reasoning tasks. However, their visual understanding behaviors remain underexplored. A fundamental question arises: to what extent do LVLMs rely on visual input, and which image regions contribute to their responses? It is non-trivial to interpret the free-form generation of LVLMs due to their complicated visual architecture (e.g., multiple encoders and multi-resolution) and variable-length outputs. In this paper, we extend existing heatmap visualization methods (e.g., iGOS++) to support LVLMs for open-ended visual question answering. We propose a method to select visually relevant tokens that reflect the relevance between generated answers and input image. Furthermore, we conduct a comprehensive analysis of state-of-the-art LVLMs on benchmarks designed to require visual information to answer. Our findings offer several insights into LVLM behavior, including the relationship between focus region and answer correctness, differences in visual attention across architectures, and the impact of LLM scale on visual understanding. The code and data are available at https://github.com/bytedance/LVLM_Interpretation.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Xiaoying Xing (6 papers)
  2. Chia-Wen Kuo (14 papers)
  3. Li Fuxin (36 papers)
  4. Yulei Niu (32 papers)
  5. Fan Chen (85 papers)
  6. Ming Li (787 papers)
  7. Ying Wu (134 papers)
  8. Longyin Wen (45 papers)
  9. Sijie Zhu (27 papers)