Here is a summary of the paper "MathFusion: Enhancing Mathematic Problem-solving of LLM through Instruction Fusion" (Pei et al., 20 Mar 2025 ):
Rationale and Problem Solved
- Problem: Existing methods to improve the math skills of LLMs often modify individual problems (like rephrasing). This doesn't effectively teach the model how different mathematical concepts and problems relate to each other.
- Goal: The paper introduces MathFusion, a new approach inspired by how humans learn math. It aims to enhance LLM mathematical reasoning by teaching them using interconnected problems, rather than isolated ones.
Data and Methods
- Fusion Strategies: MathFusion creates new training data by combining existing math problems in three ways:
- Sequential fusion: Links problems where solving one helps solve the next, mimicking dependencies.
- Parallel fusion: Presents similar problems together to strengthen understanding of core concepts.
- Conditional fusion: Creates problems that require choosing the right approach based on context, improving flexibility.
- New Dataset: Using these strategies, the researchers generated a new dataset called MathFusionQA.
- Scale: This new dataset contains 45,000 synthetic instructions (combined problems).
Models Used
- The researchers didn't create a new model from scratch. Instead, they applied the MathFusion training data to improve existing LLMs:
- DeepSeekMath-7B
- Mistral-7B
- Llama3-8B
Performance Highlights
- Models trained with MathFusion showed significant improvements on various math problem-solving tests (benchmarks).
- Accuracy increased by an average of 18.0 percentage points.
- This improvement was achieved efficiently, requiring only 45,000 additional training examples, demonstrating better data efficiency compared to traditional methods.
Implications and Applications
- Better AI Math Tutors: This technique could lead to AI that better understands and explains mathematical concepts by recognizing relationships between problems.
- Improved Scientific Assistants: LLMs enhanced with MathFusion could be more effective tools for researchers and engineers tackling complex calculations and modeling.
- Efficient AI Training: The method shows that significant performance gains can be achieved with relatively small amounts of carefully structured synthetic data, making AI training more efficient.
- Broader Reasoning Skills: While focused on math, the idea of fusing instructions based on relationships might be applicable to improving LLM reasoning in other complex domains.
In conclusion, MathFusion offers a novel and data-efficient framework for enhancing the mathematical reasoning capabilities of LLMs by training them on strategically combined problems, better reflecting how interconnected mathematical knowledge is learned.