Papers
Topics
Authors
Recent
Search
2000 character limit reached

Prioritizing the Best: Incentivizing Reliable Multimodal Reasoning by Rewarding Beyond Answer Correctness

Published 20 Apr 2026 in cs.CL | (2604.18892v1)

Abstract: Reinforcement Learning with Verifiable Rewards (RLVR) improves multimodal reasoning by rewarding verifiable final answers. Yet answer-correct trajectories may still rely on incomplete derivations, weak evidence, or statements that contradict their conclusions. This gap between answer correctness and reasoning validity, which we call reasoning-answer inconsistency, motivates trajectory supervision in multimodal RL. We compare two main approaches: reward models (RMs), and Generative Rewards (GRs). RMs are efficient and help early in training, but their gains weaken as the policy distribution shifts; GRs improve performance, but may give unstable rewards and computationally expensive. We therefore propose Groupwise Ranking Reward, which ranks verifier-passed trajectories for the same prompt in one pass and redistributes reward accordingly. Groupwise comparison better separates stronger and weaker correct trajectories with lower judge overhead than GRs. Experiments show that RLVR aggravates reasoning-answer inconsistency, while trajectory supervision alleviates it. Groupwise Ranking Reward performs best overall, improving reliability-conditioned accuracy from 47.4% to 54.7% over RLVR.

Summary

  • The paper demonstrates that trajectory-level supervision with Groupwise Ranking Reward yields simultaneous improvements in answer accuracy and reasoning faithfulness.
  • The study systematically compares reward models, generative rewards, and groupwise ranking to address the limitations of outcome-only reinforcement learning.
  • Empirical results on multimodal benchmarks show that the novel approach significantly reduces reasoning inconsistency rates while boosting reliability-conditioned accuracy.

Incentivizing Reliable Multimodal Reasoning Beyond Outcome Correctness

Problem Overview and Motivation

Standard reinforcement learning with verifiable rewards (RLVR) methods optimize multimodal LLMs (MLLMs) toward answer correctness by rewarding model-generated responses that attain the correct final answer under a rule-based verification process. However, this outcome-centric protocol neglects the quality and faithfulness of the reasoning trajectories that precede the answer. Empirical analysis demonstrates that, as RLVR training progresses, MLLMs increasingly generate answers supported by incomplete, weakly justified, or outright inconsistent chains of thought, leading to an undesirable escalation in what the authors term reasoning-answer inconsistency: responses where the answer is verifiably correct but the reasoning is misaligned or contradictory. Figure 1

Figure 1: Illustrative case of reasoning-answer inconsistency under outcome-only RLVR and a preview of Groupwise Ranking Reward’s effect on training dynamics and reward structure.

This phenomenon is particularly problematic in high-stakes applications—such as medical diagnosis or scientific computation—where reliable, interpretable reasoning is required. The core insight is that answer correctness is a poor surrogate for mechanistic reasoning quality, and optimization on outcome alone amplifies the likelihood of unfaithful reasoning trajectories. Thus, the challenge addressed is how to efficiently and effectively inject trajectory-level supervision signals into the training of MLLMs for robust, faithful multimodal reasoning.

Methods and Trajectory Supervision Formulations

To address this limitation, the paper systematically analyzes three distinct approaches to trajectory supervision that supplement standard RLVR pipelines:

  1. Reward Models (RMs): These offer direct scalar rewards, applied at either the step or trajectory level. In this work, a Vision-Language Process Reward Model (PRM) is used to assign true/false rewards to each step, aggregated to a per-trajectory score via simple averaging. This provides computationally efficient, dense supervision but is susceptible to policy distribution shift and limited generalization.
  2. Generative Rewards (GRs): A LLM-as-a-judge is prompted to evaluate each rollout independently using rubric-based natural language feedback and explicit scalar scores. This pointwise generative reward mechanism is richer and less static than fixed reward models but introduces significant judge overhead and reward noise due to repeated, isolated evaluations.
  3. Groupwise Ranking Reward: The novel contribution of the paper, this approach generalizes GRs from independent scoring to comparative evaluation across all verifier-passed rollouts from the same prompt. The judge is instructed to assign tie-aware ranking tiers instead of absolute scores, which are then mapped to normalized, zero-centered scalar rewards reflecting the relative merits of each reasoning trajectory. This enables efficient batch judging and accentuates the difference between stronger and weaker correct trajectories, thereby providing a more discriminative signal for learning. Figure 2

    Figure 2: Training dynamics for RLVR, PRM, pointwise GR, and Groupwise Ranking. Groupwise Ranking simultaneously improves answer accuracy and reliability-conditioned accuracy, while suppressing both overall and correct-but-inconsistent reasoning rates.

Experimental Results

Setup and Metrics

Extensive experiments are conducted using Qwen2.5-VL models (7B), trained with each trajectory supervision variant on the ViRL multimodal corpus and evaluated on five benchmarks spanning mathematical reasoning and general-purpose visual question answering. The primary evaluation metrics include:

  • Accuracy (Acc): Standard answer correctness.
  • Reliability-Conditioned Accuracy (RC-Acc): Correct answers with fully consistent reasoning, discounting any correct-but-inconsistent cases.
  • IncR: Overall inconsistency rate.
  • CBIR: Correct-but-inconsistent rate.

Empirical Findings

The experimental findings delineate profound differences among the evaluated reward structures:

  • Outcome-only RLVR reliably boosts standard answer accuracy (from 49.0% to 53.6%) but substantially increases the prevalence of reasoning-answer inconsistent predictions, yielding no net improvement in RC-Acc.
  • PRMs slow but do not halt the degradation of reasoning consistency, confirming their limited robustness under policy shift in multimodal tasks.
  • Pointwise GRs recover faithfulness over RLVR and offer more reliable accuracy improvements.
  • Groupwise Ranking Reward achieves dominant performance, increasing RC-Acc from baseline RLVR’s 47.4% to 54.7% and further improving answer accuracy to 55.9%. Importantly, Groupwise Ranking is the only method to consistently yield simultaneous improvements in answer accuracy and faithfulness, exhibiting reductions in both IncR and CBIR throughout training. Figure 3

    Figure 3: Scaling from 3B to 7B model size lowers baseline correct-but-inconsistent rates, but does not mitigate the upward drift under RLVR; the core failure mode is algorithmic, not architectural.

The gains from Groupwise Ranking Reward are broad-based, holding across all evaluated benchmarks and both multiple-choice and open-ended question types. Detailed ablation confirms that the specific mapping from ranks to reward has only a minor effect relative to the benefit of groupwise comparison itself.

Theoretical and Practical Implications

The central implication is that answer correctness is insufficient as a reinforcement learning objective for MLLMs, since outcome-only rewards encourage policies to exploit any path to the answer—regardless of interpretability or consistency. Trajectory supervision changes the credit assignment structure in RL, enabling the model to distinguish and prefer trajectories that are not only answer-correct but also logically valid and faithful.

From a systems perspective, Groupwise Ranking Reward offers substantial computational gains: batch-based, tie-aware ranking reduces the number of judge calls and variance in assigned rewards, making the training process more stable. Furthermore, groupwise reward assignment is more robust to fluctuations in the judge’s intrinsic scoring calibration, as intra-prompt comparison ensures a fixed reference standard for the evaluation. Figure 4

Figure 4: Ablation study across four rank-to-score mapping variants confirms negligible sensitivity, with the main gain accruing from groupwise comparison itself.

Figure 5

Figure 5: Consistency-accuracy trade-off trajectories by question type and training method; Groupwise Ranking closes the faithfulness gap for both open-ended and multiple-choice settings.

This trajectory-level supervision framework is readily extensible to other RL pipelines optimizing for faithful multistep reasoning, and is especially pertinent for multimodal domains where the alignment between visual/textual evidence and final predictions cannot be assumed.

Limitations and Future Directions

The scope of this analysis is confined to tasks for which final answer correctness is verifiable and trajectory-level pathologies are well-defined. Applications involving subjective, creative, or unstructured outputs may require new consistency metrics and reward paradigms. Additionally, while groupwise comparison is shown to be highly effective, its reliability is contingent on the robustness of the judge model to adversarial or out-of-distribution reasoning chains. Figure 6

Figure 6: Models trained on text-only arithmetic (GSM8K) maintain low reasoning inconsistency, while multimodal ViRL training achieves stronger accuracy but at increased risk of spurious reasoning.

Open directions include generalizing ranking-based reward frameworks to non-verifiable or open-ended multimodal tasks, investigating judge robustness against prompt distribution shift, and designing scalable, adaptive judge policies that can better audit and incentivize faithful reasoning under adversarial conditions.

Conclusion

This study demonstrates that trajectory-level reward assignment, especially via groupwise ranking, is critical for closing the gap between answer correctness and reliable reasoning in MLLMs. By redistributing reward among verifier-passed rollouts according to relative reasoning quality, Groupwise Ranking Reward offers both theoretical and empirical advances—improving the training signal’s faithfulness and the ultimate trustworthiness of model-generated rationales. This line of work motivates a reorientation of RL training objectives in MLLMs toward comprehensive, consistency-aware supervision paradigms for high-assurance reasoning.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 0 likes about this paper.