Papers
Topics
Authors
Recent
AI Research Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 67 tok/s
Gemini 2.5 Pro 36 tok/s Pro
GPT-5 Medium 16 tok/s Pro
GPT-5 High 18 tok/s Pro
GPT-4o 66 tok/s Pro
Kimi K2 170 tok/s Pro
GPT OSS 120B 440 tok/s Pro
Claude Sonnet 4 36 tok/s Pro
2000 character limit reached

Visual-TableQA: Open-Domain Benchmark for Reasoning over Table Images (2509.07966v1)

Published 9 Sep 2025 in cs.CV and cs.CL

Abstract: Visual reasoning over structured data such as tables is a critical capability for modern vision-LLMs (VLMs), yet current benchmarks remain limited in scale, diversity, or reasoning depth, especially when it comes to rendered table images. Addressing this gap, we introduce Visual-TableQA, a large-scale, open-domain multimodal dataset specifically designed to evaluate and enhance visual reasoning over complex tabular data. Our generation pipeline is modular, scalable, and fully autonomous, involving multiple reasoning LLMs collaborating across distinct roles: generation, validation, and inspiration. Visual-TableQA comprises 2.5k richly structured LaTeX-rendered tables and 6k reasoning-intensive QA pairs, all produced at a cost of under USD 100. To promote diversity and creativity, our pipeline performs multi-model collaborative data generation via cross-model prompting ('inspiration') and LLM-jury filtering. Stronger models seed layouts and topics that weaker models elaborate, collectively distilling diverse reasoning patterns and visual structures into the dataset. Empirical results show that models fine-tuned on Visual-TableQA generalize robustly to external benchmarks, outperforming several proprietary models despite the dataset's synthetic nature. The full pipeline and resources are publicly available at https://github.com/AI-4-Everyone/Visual-TableQA.

Summary

  • The paper introduces a scalable, open-domain dataset for reasoning over complex table images using a modular LaTeX generation pipeline.
  • The methodology integrates cross-model inspiration, automated QA generation, and multi-stage validation to ensure high-quality, diverse data.
  • Experimental findings show enhanced transferability and improved performance on both visual and textual reasoning tasks across model types.

Visual-TableQA: Open-Domain Benchmark for Reasoning over Table Images

Motivation and Context

Visual reasoning over structured data, particularly rendered table images, remains a significant challenge for vision-LLMs (VLMs). Existing benchmarks are either limited in scale, diversity, or reasoning depth, and most focus on textual representations or domain-specific layouts, failing to capture the complexity of real-world tabular data. Visual-TableQA addresses these gaps by introducing a large-scale, open-domain multimodal dataset specifically designed to evaluate and enhance visual reasoning over complex table images. The dataset leverages a modular, scalable, and fully autonomous generation pipeline, integrating multiple LLMs for generation, validation, and inspiration, and is produced at minimal cost.

Dataset Construction and Pipeline

Visual-TableQA comprises 2,500 richly structured LaTeX-rendered tables and 6,000 reasoning-intensive QA pairs. The pipeline is built around the principles of Code-as-Intermediary Translation (CIT), using LaTeX as an intermediate representation to efficiently generate complex tables and minimize annotation costs.

The pipeline consists of several stages:

  1. Seed Collection: Diverse table layouts and topics are collected from various sources, including scientific journals and financial reports. Initial table images are converted to LaTeX using a visual LLM (VLM-0), and 5,000 topic prompts are generated by an LLM.
  2. Table Generation: For each iteration, an LLM receives a table sample and three topics, generating three new LaTeX tables inspired by the input but with substantial layout and data variations. The LaTeX code is compiled and visually inspected for quality.
  3. Cross-Model Inspiration: Generated tables are iteratively added to the inspiration pool, enabling cross-model prompting and fostering layout diversity through collaborative synthesis.
  4. QA Generation: Another LLM generates three reasoning-intensive QA pairs per table, focusing on multi-step reasoning, pattern recognition, and symbolic interpretation.
  5. Quality Assurance: A jury of high-performing LLMs evaluates each table and QA pair for validity, coherence, and grounding, using the ROSCOE reasoning score for step-by-step rationale assessment. Human annotators further validate a subset of QA pairs. Figure 1

    Figure 1: Full pipeline architecture of Visual-TableQA, illustrating the multi-stage, multi-model collaborative generation and validation process.

Dataset Properties: Diversity and Complexity

Visual-TableQA emphasizes structural reasoning over domain knowledge, coupling rendered table images with visually grounded reasoning tasks. Tables feature complex layouts, multirow cells, integrated diagrams, and color encoding, requiring models to interpret visual cues and perform multi-step logical analysis.

Topic diversity is ensured by clustering 5,000 topics using UMAP and K-Means, resulting in balanced semantic coverage. Layout diversity is achieved through iterative cross-model inspiration, producing a wide range of table structures from limited seeds. Figure 2

Figure 2: Cumulative topic coverage curve, demonstrating uniform distribution of topics across clusters.

Figure 3

Figure 3: 2D projection of 5,000 topics, with semantic clusters and representative topics illustrating diversity.

Figure 4

Figure 4

Figure 4: Initial layout seeds used for table generation, serving as inspiration for subsequent iterations.

Experimental Evaluation

Visual-TableQA is benchmarked against ChartQA, ReachQA, and MATH-Vision, covering recognition-heavy, balanced, and reasoning-heavy tasks, respectively. A broad spectrum of VLMs, both proprietary and open-source, are evaluated using a unified prompt and majority-vote LLM jury.

Key findings include:

  • Correlation with External Benchmarks: Model rankings on Visual-TableQA strongly correlate with ReachQA, indicating effective balance between visual recognition and reasoning. Weaker correlation with ChartQA and MATH-Vision highlights Visual-TableQA's unique position as a comprehensive visual reasoning benchmark. Figure 5

    Figure 5: Correlation of model rankings on Visual-TableQA with ChartQA, ReachQA, and MATH-Vision using Spearman’s ρ\rho and Kendall’s τ\tau.

  • Transferability: Fine-tuning on Visual-TableQA yields substantial generalization gains on external benchmarks, notably improving reasoning performance on ReachQA and MATH-Vision. This transferability is not reciprocal; fine-tuning on ReachQA does not yield similar gains on Visual-TableQA.
  • Proprietary vs. Open-Source Models: Proprietary models (e.g., Claude 3.5 Sonnet) outperform open-source models on average, but Visual-TableQA fine-tuning narrows this gap, with open-source models surpassing several proprietary baselines.
  • Image vs. Textual Input: Models perform +6.26%+6.26\% better on the textual (LaTeX code) variant of Visual-TableQA, underscoring the added challenge of visual reasoning over rendered images.

Error Analysis and Quality Metrics

A detailed error taxonomy for Qwen2.5-VL-7B-Instruct reveals that fine-tuning on Visual-TableQA shifts errors toward incoherence, with reductions in hallucination and partial data extraction. This suggests more consistent reasoning patterns post-finetuning, but highlights the need for targeted synthetic supervision to address specific error types. Figure 6

Figure 6

Figure 6: Histogram of errors for Qwen2.5-vl-7b, illustrating error distribution before and after fine-tuning.

ROSCOE metrics indicate near-perfect semantic adequacy, low redundancy, high logical inference, and strong grammaticality, supporting the overall quality of generated rationales.

Practical Implications and Future Directions

Visual-TableQA's modular, scalable pipeline enables cost-efficient, high-quality dataset generation, facilitating knowledge distillation and transfer learning for multimodal reasoning tasks. The collaborative cross-model inspiration mechanism fosters diversity and creativity, making the dataset particularly effective for training models with transferable reasoning skills.

The dataset's strong empirical results and transferability suggest its utility as a general-purpose benchmark for visual reasoning over structured data. However, limitations remain in LaTeX expressiveness for complex images and the reliability of automated evaluation metrics. Future work should focus on developing robust bidirectional image-to-text encoding systems and hybrid evaluation strategies combining automated and human verification.

Conclusion

Visual-TableQA establishes a new standard for open-domain visual reasoning benchmarks over table images, offering unprecedented diversity, reasoning depth, and scalability. Its pipeline demonstrates that synthetic datasets, when carefully constructed and validated, can effectively challenge and improve state-of-the-art VLMs. The dataset's transferability and modularity position it as a valuable resource for advancing multimodal reasoning and bridging the performance gap between open-source and proprietary models.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Github Logo Streamline Icon: https://streamlinehq.com
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 9 posts and received 26 likes.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube