The paper "Open Eyes, Then Reason: Fine-grained Visual Mathematical Understanding in MLLMs" addresses the challenges faced by multimodal LLMs (MLLMs) in solving mathematical problems that require a nuanced comprehension of visual data. It highlights the limitations of existing MLLMs, such as GPT-4o, which exhibit a high error rate (approximately 70%) when identifying geometric primitives—a critical aspect of visual mathematical reasoning.
The paper identifies that the current underperformance can be largely attributed to inadequate perception of geometric primitives during image-level contrastive pre-training processes, such as those used by CLIP. Current attempts to address these issues often focus on increasing the scale of mathematical visual instruction datasets and strengthening the backbone LLMs. However, these efforts sometimes neglect persistent errors in visual recognition, which have a significant negative correlation with problem-solving performance.
To address these limitations, the paper proposes a novel approach called SVE-Math (Selective Vision-Enhanced Mathematical MLLM). This framework includes:
- A Geometric-Grounded Vision Encoder: A specialized encoder tasked with recognizing and accurately grounding geometric primitives. This targets the root cause of visual misrecognition by enhancing the model's perception capabilities.
- A Feature Router: This dynamic mechanism adjusts the contribution of hierarchical visual feature maps, ensuring that the MLLM receives pertinent visual information without redundant cues. The feature router creates visual soft prompts that are tailored to the LLM's needs, improving the reasoning process.
The proposed SVE-Math model demonstrates significant improvements in handling mathematical visual reasoning tasks. In experiments, SVE-Math-Qwen2.5-7B, a model using the proposed framework, outperforms other 7B parameter models by a substantial margin—15% on the MathVerse benchmark—and achieves performance compatibility with advanced models like GPT-4V on the MathVista dataset. Using smaller training datasets, SVE-Math also achieves competitive results on GeoQA, approaching the efficacy of models trained on larger datasets.
This paper underscores the importance of integrating fine-grained visual understanding and adaptive visual cue processing into multimodal LLMs. By shifting the focus towards enhancing visual perception rather than merely scaling data or computational complexity, the paper offers a promising direction for future research in the development of capable visual mathematical reasoning systems within MLLMs. The authors provide the implementation of SVE-Math on their GitHub repository for further exploration and development by the community.