Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

MathFusion: Enhancing Mathematical Problem-solving of LLM through Instruction Fusion (2503.16212v2)

Published 20 Mar 2025 in cs.CL and cs.AI

Abstract: LLMs have shown impressive progress in mathematical reasoning. While data augmentation is promising to enhance mathematical problem-solving ability, current approaches are predominantly limited to instance-level modifications-such as rephrasing or generating syntactic variations-which fail to capture and leverage the intrinsic relational structures inherent in mathematical knowledge. Inspired by human learning processes, where mathematical proficiency develops through systematic exposure to interconnected concepts, we introduce MathFusion, a novel framework that enhances mathematical reasoning through cross-problem instruction synthesis. MathFusion implements this through three fusion strategies: (1) sequential fusion, which chains related problems to model solution dependencies; (2) parallel fusion, which combines analogous problems to reinforce conceptual understanding; and (3) conditional fusion, which creates context-aware selective problems to enhance reasoning flexibility. By applying these strategies, we generate a new dataset, \textbf{MathFusionQA}, followed by fine-tuning models (DeepSeekMath-7B, Mistral-7B, Llama3-8B) on it. Experimental results demonstrate that MathFusion achieves substantial improvements in mathematical reasoning while maintaining high data efficiency, boosting performance by 18.0 points in accuracy across diverse benchmarks while requiring only 45K additional synthetic instructions, representing a substantial improvement over traditional single-instruction approaches. Our datasets, models, and code are publicly available at https://github.com/QizhiPei/mathfusion.

Here is a summary of the paper "MathFusion: Enhancing Mathematic Problem-solving of LLM through Instruction Fusion" (Pei et al., 20 Mar 2025 ):

Rationale and Problem Solved

  • Problem: Existing methods to improve the math skills of LLMs often modify individual problems (like rephrasing). This doesn't effectively teach the model how different mathematical concepts and problems relate to each other.
  • Goal: The paper introduces MathFusion, a new approach inspired by how humans learn math. It aims to enhance LLM mathematical reasoning by teaching them using interconnected problems, rather than isolated ones.

Data and Methods

  • Fusion Strategies: MathFusion creates new training data by combining existing math problems in three ways:
    • Sequential fusion: Links problems where solving one helps solve the next, mimicking dependencies.
    • Parallel fusion: Presents similar problems together to strengthen understanding of core concepts.
    • Conditional fusion: Creates problems that require choosing the right approach based on context, improving flexibility.
  • New Dataset: Using these strategies, the researchers generated a new dataset called MathFusionQA.
  • Scale: This new dataset contains 45,000 synthetic instructions (combined problems).

Models Used

  • The researchers didn't create a new model from scratch. Instead, they applied the MathFusion training data to improve existing LLMs:
    • DeepSeekMath-7B
    • Mistral-7B
    • Llama3-8B

Performance Highlights

  • Models trained with MathFusion showed significant improvements on various math problem-solving tests (benchmarks).
  • Accuracy increased by an average of 18.0 percentage points.
  • This improvement was achieved efficiently, requiring only 45,000 additional training examples, demonstrating better data efficiency compared to traditional methods.

Implications and Applications

  • Better AI Math Tutors: This technique could lead to AI that better understands and explains mathematical concepts by recognizing relationships between problems.
  • Improved Scientific Assistants: LLMs enhanced with MathFusion could be more effective tools for researchers and engineers tackling complex calculations and modeling.
  • Efficient AI Training: The method shows that significant performance gains can be achieved with relatively small amounts of carefully structured synthetic data, making AI training more efficient.
  • Broader Reasoning Skills: While focused on math, the idea of fusing instructions based on relationships might be applicable to improving LLM reasoning in other complex domains.

In conclusion, MathFusion offers a novel and data-efficient framework for enhancing the mathematical reasoning capabilities of LLMs by training them on strategically combined problems, better reflecting how interconnected mathematical knowledge is learned.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Qizhi Pei (17 papers)
  2. Lijun Wu (113 papers)
  3. Zhuoshi Pan (9 papers)
  4. Yu Li (377 papers)
  5. Honglin Lin (9 papers)
  6. Chenlin Ming (6 papers)
  7. Xin Gao (208 papers)
  8. Conghui He (114 papers)
  9. Rui Yan (250 papers)