Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 83 tok/s
Gemini 2.5 Pro 54 tok/s Pro
GPT-5 Medium 21 tok/s Pro
GPT-5 High 20 tok/s Pro
GPT-4o 103 tok/s Pro
Kimi K2 205 tok/s Pro
GPT OSS 120B 456 tok/s Pro
Claude Sonnet 4 35 tok/s Pro
2000 character limit reached

Perception-R1: Enhancing Multimodal Visual Perception

Updated 21 September 2025
  • Perception-R1 is a reinforcement learning framework that introduces a novel visual perception reward to compel explicit comprehension of visual details before reasoning.
  • Its methodology employs detailed chain-of-thought visual annotations and group relative policy optimization to validate intermediate perceptual steps with precision.
  • Experimental benchmarks demonstrate that Perception-R1 outperforms prior approaches on multimodal reasoning tasks, achieving statistically significant improvements with fewer training samples.

Perception-R1 designates a reinforcement learning framework that aims to directly optimize and enhance the perceptual abilities of multimodal LLMs (MLLMs) through explicit task-level reward signals. Its primary distinguishing feature is the introduction of a visual perception reward that compels the model not only to produce accurate final answers on multimodal reasoning problems, but to demonstrate and verify explicit comprehension of visual content as an essential precursor to reasoning. This construct addresses the limitation—identified in prior RL with Verifiable Reward (RLVR) approaches—where answer accuracy alone does not suffice to ensure improved perception, especially in settings with sparse or weak reward signals. As such, Perception-R1 sits at the intersection of reinforcement learning, visual-LLM post-training, and cognitive modeling of perception in complex reasoning tasks.

1. Motivation and Conceptual Foundations

Perception-R1 is motivated by empirical findings that standard RLVR post-training of MLLMs, which rewards correct answers and output formatting, often fails to materially improve the model's multimodal perception capabilities. The fundamental insight is that multimodal reasoning relies on an accurate perception phase: answer correctness is a necessary but not sufficient metric, since it is possible for a model to reach the right answer using erroneous or hallucinated interpretations of the input image or diagram.

To overcome this, Perception-R1 decouples visual perception from answer accuracy in its reward formulation. The model is rewarded for accurately describing the critical elements of the visual input prior to producing the answer. Visual perception rewards are systematically extracted from detailed chain-of-thought (CoT) trajectories—these contain atomic visual annotations (such as geometric relationships, object properties, or positional cues) that represent accurate intermediate perceptual steps necessary for solving the problem. During post-training, adherence to these visual facts is evaluated and incorporated into the RL signal.

2. Methodology and Reward Mechanism

The Perception-R1 methodology comprises several operational stages:

  • Construction of Visual Annotations: CoT trajectories produced by a competent MLLM on a multimodal math dataset are analyzed by a text-only LLM, which extracts a set V={vj}j=1m\mathcal{V} = \{v_j\}_{j=1}^m of granular visual annotations corresponding to non-redundant visual facts (such as "segment GEDF\overline{GE}\perp\overline{DF}" or numerical attributes "GE = 10").
  • Response Evaluation via Judging LLM: During RLVR post-training, a generated response yiy_i is evaluated for visual perception quality. A judging LLM is prompted to verify, for each annotation vjv_j, whether yiy_i explicitly captures the visual detail (i.e., oi,j=Φ(yi,vj){0,1}o_{i,j} = \Phi(y_i, v_j) \in \{0, 1\}, with $1$ for a match).
  • Composite Reward Assignment: The visual perception reward rv(yi,V)r_v(y_i, \mathcal{V}) is computed as

rv(yi,V)=1mj=1moi,jr_v(y_i, \mathcal{V}) = \frac{1}{m} \sum_{j=1}^m o_{i,j}

and contributes to the total reward for each sampled output according to

r(yi,a,V)=αrf(yi)+βra(yi,a)+γrv(yi,V)+rp(yi)r(y_i, a, \mathcal{V}) = \alpha \cdot r_f(y_i) + \beta \cdot r_a(y_i, a) + \gamma \cdot r_v(y_i, \mathcal{V}) + r_p(y_i)

where rfr_f is a format reward, rar_a is answer correctness, γ\gamma weights the perception reward, and rpr_p penalizes repetitions.

  • Policy Optimization via GRPO: The training employs Group Relative Policy Optimization (GRPO), a PPO variant. For each batch, a group of GG model outputs is sampled; normalized advantage estimators

A^i=r(yi,a,V)mean()std()\hat{A}_i = \frac{r(y_i, a, \mathcal{V}) - \mathrm{mean}(\dots)}{\mathrm{std}(\dots)}

are computed and used to update the policy under a clipped ratio constraint and a KL regularization to remain close to the reference policy.

This reward design aligns the model’s optimization trajectory with accurate perception as well as correct problem-solving.

3. Experimental Validation and Benchmarks

Perception-R1 is validated on representative multimodal math reasoning benchmarks, including MathVista, MathVerse (with subsets like "Vision Only"), MathVision, and WeMath. Training is conducted on a subset of 1,442 Geometry3K samples with valid, high-quality CoT and visual annotations; this data efficiency is notable compared to the hundreds of thousands of samples used by competitors.

Performance metrics include benchmark accuracy and statistical measures of significance. Perception-R1 achieves state-of-the-art results:

  • Superior average scores on MathVista and MathVerse relative to prior art.
  • On the "Vision Only" subset of MathVerse, Perception-R1 demonstrably outperforms both the standard base MLLM and RLVR-trained variants omitting the visual reward.
  • McNemar's test on visual perception error reductions reports p=0.04p=0.04—indicating statistically significant improvements.
  • The improvement is achieved with a minimal data regime, illustrating the reward’s efficiency and informativeness.

Ablation studies confirm that removing the visual perception reward or repetition penalty materially degrades model performance, highlighting the necessity of these components, while answer-only rewards are insufficient to enhance perceptual fidelity.

4. Comparison with Prior and Contemporary Methods

Contrasted against RLVR, which combines answer and format rewards, Perception-R1 uniquely employs visual perception-derived signals. RLVR systems frequently exhibit reward sparsity, permitting the model to reach correct answers after hallucinating, ignoring, or fabricating critical perceptual facts. By introducing per-annotation rewards, Perception-R1 mitigates this failure mode, explicitly incentivizing the describe-then-solve strategy adopted by expert human solvers.

Relative to chain-of-thought reasoning methods that optimize for answer accuracy and rationalized explanations, Perception-R1 ensures that the explanation phase is anchored to genuine visual inspection, not post hoc rationalization. This precludes a class of reward hacking behaviors where the model optimizes for plausible reasoning chains that sidestep perception mistakes.

Perception-R1's approach differs from strategies relying on end-to-end system performance metrics or global reward signals which can be too delayed or aggregated to provide an effective learning signal for perception sub-tasks.

5. Implications for Model Architecture and Training

Perception-R1's findings suggest several architectural and procedural implications for MLLM and LVLM training pipelines:

  • The decoupling of perception and reasoning allows for targeted enhancement of each stage, suggesting that modular training objectives can yield more robust, interpretable, and trustworthy models.
  • High-quality intermediate visual annotations can serve as an efficient and scalable supervision source where strong answer supervision is either sparse or uninformative.
  • The explicit judgment of perception accuracy requires reliable judging LLMs; as such, further improvements in validating visual annotations and mitigating judging bias are necessary for scaling this approach.
  • The methodology is readily extensible to non-math domains, conditional on developing domain-appropriate visual annotation extraction and verification mechanisms.

Potential limitations include the dependency on the quality of the base model's CoT trajectories for extracting visual annotations and on the accuracy of the judging LLM.

6. Broader Impact and Future Directions

Perception-R1 substantiates the hypothesis that reward functions in RL post-training for MLLMs must encompass more than answer accuracy to achieve genuine multimodal comprehension. By enforcing pre-answer perception verification, it elevates the robustness of multimodal reasoning, yielding models that not only answer correctly but also "see" and interpret accurately.

Potential extensions include:

  • Applying visual perception reward schemas to other domains (e.g., science diagrams, technical document understanding, medical imaging) where correct perception is necessary before reasoning.
  • Developing richer, perhaps soft-valued, visual reward functions to capture partial perceptions or graded confidence levels.
  • Integrating more advanced or domain-adapted judging LLMs to reduce the possibility of introducing verification biases.
  • Investigating methods for automatic extraction of perceptual annotations from non-CoT trajectories, expanding the applicability beyond math-intensive contexts.

Ongoing research will further probe the link between explicit perceptual supervision, architectural modularity, and the generalizability of MLLM reasoning across diverse multimodal applications.

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Perception-R1.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube