Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 147 tok/s
Gemini 2.5 Pro 40 tok/s Pro
GPT-5 Medium 28 tok/s Pro
GPT-5 High 24 tok/s Pro
GPT-4o 58 tok/s Pro
Kimi K2 201 tok/s Pro
GPT OSS 120B 434 tok/s Pro
Claude Sonnet 4.5 38 tok/s Pro
2000 character limit reached

Effective Training Data Synthesis for Improving MLLM Chart Understanding (2508.06492v1)

Published 8 Aug 2025 in cs.CV and cs.CL

Abstract: Being able to effectively read scientific plots, or chart understanding, is a central part toward building effective agents for science. However, existing multimodal LLMs (MLLMs), especially open-source ones, are still falling behind with a typical success rate of 30%-50% on challenging benchmarks. Previous studies on fine-tuning MLLMs with synthetic charts are often restricted by their inadequate similarity to the real charts, which could compromise model training and performance on complex real-world charts. In this study, we show that modularizing chart generation and diversifying visual details improves chart understanding capabilities. In particular, we design a five-step data synthesis pipeline, where we separate data and function creation for single plot generation, condition the generation of later subplots on earlier ones for multi-subplot figures, visually diversify the generated figures, filter out low quality data, and finally generate the question-answer (QA) pairs with GPT-4o. This approach allows us to streamline the generation of fine-tuning datasets and introduce the effective chart dataset (ECD), which contains 10k+ chart images and 300k+ QA pairs, covering 25 topics and featuring 250+ chart type combinations with high visual complexity. We show that ECD consistently improves the performance of various MLLMs on a range of real-world and synthetic test sets. Code, data and models are available at: https://github.com/yuweiyang-anu/ECD.

Summary

  • The paper introduces the Effective Chart Dataset (ECD) with over 10,000 chart images and 321,544 QA pairs to enhance MLLM chart interpretation.
  • The methodology employs a modular five-step synthesis pipeline to increase data diversity, visual complexity, and semantic coherence.
  • Experiments reveal that models fine-tuned with ECD outperform previous datasets on benchmarks, validating its effectiveness in chart understanding.

Summary of "Effective Training Data Synthesis for Improving MLLM Chart Understanding"

This paper presents a novel methodology to enhance chart understanding in multimodal LLMs (MLLMs) through effective training data synthesis. The authors introduce the Effective Chart Dataset (ECD), a high-quality synthetic dataset designed to bridge the gap between synthetic and real-world chart understanding.

Introduction and Motivation

Chart understanding is crucial for building effective AI agents capable of interpreting scientific data. Current MLLMs, while proficient in text-based tasks, struggle with chart understanding, exhibiting low success rates on complex benchmarks. This challenge is exacerbated by the limitations of existing synthetic datasets, which often lack the fidelity and diversity of real-world charts.

The ECD is designed to address these shortcomings using a five-step data synthesis pipeline that focuses on improving data quality and diversity. This pipeline modularizes chart generation processes and introduces visual diversity in the generated datasets, setting a new standard for chart training sets.

Data Synthesis Pipeline

Single Plot Generation

The process begins with generating single plots using pre-defined themes and chart functions in Python. GPT-4o is tasked with producing data and textual elements like titles and labels, ensuring a rich distribution of data while maintaining semantic coherence.

Combined Subplots Generation

A conditional generation strategy is used for creating multi-subplot charts. Each subplot generation is conditioned on prior subplots to maintain thematic and semantic coherence, akin to human-designed scientific figures.

Chart Image Diversification

To overcome stylistic limitations, the charts undergo a diversification process which modifies visual elements like annotations, color schemes, and layout adjustments. This step ensures that the charts exhibit higher visual complexity and variation.

Filtering Low-quality Charts

Charts are evaluated based on visual clarity and semantic coherence using specialized GPT-4o prompts. This filtering step ensures that only high-quality chart images, which meet or exceed average quality standards, are included in the final dataset.

QA Pairs Generation

Finally, GPT-4o generates descriptive and reasoning-based question-answer pairs for the charts. The QA pairs are rigorously filtered to ensure high confidence and accuracy, maintaining the dataset's integrity as a training resource. Figure 1

Figure 1: Overview of the effective chart dataset (ECD). It includes a diverse range of chart images and QA pairs.

Effective Chart Dataset (ECD)

ECD includes over 10,000 chart images and 321,544 QA pairs spanning 29 chart types and 252 subplot combinations. It significantly outperforms existing datasets in terms of visual diversity and complexity, as evidenced by its superior performance on benchmarks like CharXiv and others.

Experiments and Results

The ECD was tested on several benchmarks, and models fine-tuned with ECD consistently outperformed those fine-tuned with other datasets. Improvements were observed across various metrics, highlighting the effectiveness of ECD in enhancing chart understanding capabilities in MLLMs.

Comparison with Other Datasets

Comparisons against other chart datasets demonstrated ECD's superior ability to enhance model performance across multiple test sets, indicating its potential as a standard resource for training MLLMs on chart understanding tasks. Figure 2

Figure 2: Our chart generation pipeline consists of five key steps.

Ablation Studies

Ablation studies confirmed the importance of diverse chart types and themes, and the inclusion of both descriptive and reasoning QA pairs in improving model performance. The chart types, themes, and QA diversification strategies were critical to achieving superior results.

Conclusion

The paper introduces and validates ECD, demonstrating its efficacy in improving chart understanding for MLLMs. By addressing the limitations of existing training data through this novel synthesis approach, the authors make a significant contribution to the field of AI-driven chart comprehension. Future work may explore further expansions of ECD and application-specific fine-tuning to unlock even greater potential in chart understanding AI models.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 2 tweets and received 17 likes.

Upgrade to Pro to view all of the tweets about this paper: