Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

ChartMuseum: Testing Visual Reasoning Capabilities of Large Vision-Language Models (2505.13444v1)

Published 19 May 2025 in cs.CL and cs.CV

Abstract: Chart understanding presents a unique challenge for large vision-LLMs (LVLMs), as it requires the integration of sophisticated textual and visual reasoning capabilities. However, current LVLMs exhibit a notable imbalance between these skills, falling short on visual reasoning that is difficult to perform in text. We conduct a case study using a synthetic dataset solvable only through visual reasoning and show that model performance degrades significantly with increasing visual complexity, while human performance remains robust. We then introduce ChartMuseum, a new Chart Question Answering (QA) benchmark containing 1,162 expert-annotated questions spanning multiple reasoning types, curated from real-world charts across 184 sources, specifically built to evaluate complex visual and textual reasoning. Unlike prior chart understanding benchmarks -- where frontier models perform similarly and near saturation -- our benchmark exposes a substantial gap between model and human performance, while effectively differentiating model capabilities: although humans achieve 93% accuracy, the best-performing model Gemini-2.5-Pro attains only 63.0%, and the leading open-source LVLM Qwen2.5-VL-72B-Instruct achieves only 38.5%. Moreover, on questions requiring primarily visual reasoning, all models experience a 35%-55% performance drop from text-reasoning-heavy question performance. Lastly, our qualitative error analysis reveals specific categories of visual reasoning that are challenging for current LVLMs.

Overview of ChartMuseum: Visual Reasoning in Large Vision-LLMs

The paper, "ChartMuseum: Testing Visual Reasoning Capabilities of Large Vision-LLMs," discusses the development and evaluation of a new benchmark for assessing the visual reasoning capabilities of large vision-LLMs (LVLMs). The primary focus is the gap between textual and visual reasoning skills within these models, highlighting that visual reasoning remains a challenging area for LVLMs.

ChartMuseum is introduced as a benchmark specifically curated to evaluate complex visual and textual reasoning via a diverse set of charts and questions. The benchmark comprises 1,162 expert-annotated questions from real-world charts sourced across 184 domains. Importantly, all questions in ChartMuseum were created by human annotators without machine assistance, ensuring high-quality, realistic assessments. The benchmark reveals that even the best-performing models fall significantly short of human-level accuracy, with humans achieving 93% accuracy, while the leading model, Gemini-2.5-Pro, achieves only 63.0%.

Key Findings and Contributions

  • Visual Reasoning Limitation: The paper identifies a major performance drop in LVLMs for questions requiring visual reasoning compared to those needing textual reasoning. Models like GPT-4.1 experience a 50% absolute accuracy decline from textual to visual tasks. This underlines a persistent area for improvement in LVLM development.
  • ChartMuseum Benchmark: The benchmark sets itself apart by exposing performance disparities across models and emphasizing the difficulty in visual reasoning. Its design prevents the saturation seen in previous benchmarks, offering a more rigorous evaluation of model capabilities.
  • Synthetic Dataset Experiment: Through experiments with synthetic charts requiring visual reasoning, the paper demonstrates model performance degradation as chart complexity increases, contrasting with consistent human accuracy. This reinforces the need for benchmarks testing beyond basic text extraction tasks.
  • Error Analysis: The error analysis categorizes common shortcomings in model visual reasoning tasks, such as symbol selection, visual comparison, trajectory tracking, and x/y value identification. Models tend to over-rely on textual strategies even when visual reasoning is more efficient.

Implications and Future Directions

This research has implications for future LVLM design, especially for applications requiring robust visual reasoning capabilities. The large performance gap between models and humans suggests a need for improved vision encoders and alignment of visual features within models. Insights drawn from ChartMuseum could inform more effective training paradigms and the development of specialized architectures to enhance visual reasoning.

In conclusion, ChartMuseum sets a new standard for evaluating LVLMs by focusing on the nuanced challenges of visual reasoning. The benchmark can guide future advancements in multimodal AI, driving research to develop models that approach human-level performance in both textual and visual reasoning tasks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (15)
  1. Liyan Tang (12 papers)
  2. Grace Kim (5 papers)
  3. Xinyu Zhao (54 papers)
  4. Thom Lake (5 papers)
  5. Wenxuan Ding (14 papers)
  6. Fangcong Yin (8 papers)
  7. Prasann Singhal (7 papers)
  8. Manya Wadhwa (8 papers)
  9. Zeyu Leo Liu (5 papers)
  10. Zayne Sprague (10 papers)
  11. Ramya Namuduri (2 papers)
  12. Bodun Hu (6 papers)
  13. Juan Diego Rodriguez (12 papers)
  14. Puyuan Peng (21 papers)
  15. Greg Durrett (117 papers)
Youtube Logo Streamline Icon: https://streamlinehq.com