Papers
Topics
Authors
Recent
Search
2000 character limit reached

Pass@k Training for Adaptively Balancing Exploration and Exploitation of Large Reasoning Models

Published 14 Aug 2025 in cs.LG, cs.AI, and cs.CL | (2508.10751v1)

Abstract: Reinforcement learning with verifiable rewards (RLVR), which typically adopts Pass@1 as the reward, has faced the issues in balancing exploration and exploitation, causing policies to prefer conservative actions, converging to a local optimum. Identifying an appropriate reward metric is therefore crucial. Regarding the prior work, although Pass@k has been used in evaluation, its connection to LLM exploration ability in RLVR remains largely overlooked. To investigate this, we first use Pass@k as the reward to train the policy model (i.e., $\textbf{Pass@k Training}$), and observe the improvement on its exploration ability. Next, we derive an analytical solution for the advantage of Pass@k Training, leading to an efficient and effective process. Building on this, our analysis reveals that exploration and exploitation are not inherently conflicting objectives, while they can mutually enhance each other. Moreover, Pass@k Training with analytical derivation essentially involves directly designing the advantage function. Inspired by this, we preliminarily explore the advantage design for RLVR, showing promising results and highlighting a potential future direction.

Summary

  • The paper introduces Pass@k Training, leveraging the Pass@k reward metric to enhance exploration in large reasoning models without degrading Pass@1 performance.
  • It employs both full sampling and bootstrap sampling techniques to efficiently generate diverse responses while reducing computational costs and variability.
  • Analytical derivation of advantage functions and implicit reward designs stabilize training and improve generalization across various tasks and domains.

Pass@k Training for Adaptively Balancing Exploration and Exploitation of Large Reasoning Models

Introduction

The paper introduces a novel approach called Pass@k Training, designed to enhance the reinforcement learning with verifiable rewards (RLVR) for large reasoning models (LRMs). Traditional RLVR methods often focus on the Pass@1 metric, which optimizes for a single, most confident response, leading to imbalanced exploration and exploitation. The proposed Pass@k Training uses the Pass@k metric as a reward to improve exploration capabilities, allowing the model to generate multiple diverse responses, potentially capturing a wider range of solutions.

Pass@k as Reward in RLVR Training

The Pass@k metric evaluates whether a model can generate a successful response within kk attempts. By integrating this into the RLVR process, the paper aims to address the inherent limitations of Pass@1 Training, where exploration is often sacrificed for exploitation.

Full Sampling Implementation

The basic implementation of Pass@k Training involves rolling out NrolloutN_{\text{rollout}} responses, separating them into groups, and assigning rewards based on the maximum success among the sampled responses. This method demonstrated significant improvements in exploration capabilities without degrading Pass@1 performance (Figure 1). Figure 1

Figure 1

Figure 1: Training progress of Pass@1 Training and Pass@k Training with Full Sampling on baseline setting.

Efficient Training with Bootstrap Sampling

To optimize the computational efficiency, bootstrap sampling is introduced. This method constructs multiple groups from the same rollout responses, reducing the required computational resources while maintaining performance. Bootstrap sampling achieved higher training efficiency compared to full sampling (Figure 2). Figure 2

Figure 2

Figure 2: Training progress of Pass@1 Training and Pass@k Training with Bootstrap Sampling under various $N_\text{rollout$.

Analytical Derivation

An analytical approach to derive the advantage values is presented, reducing variability introduced by sampling and further enhancing the efficiency of Pass@k Training (Figure 3). This method allows for a more stable training process, contributing to continuous performance improvements in exploration tasks. Figure 3

Figure 3

Figure 3: Training progress of Pass@1 Training and Pass@k Training with Analytical Derivation and Bootstrap Sampling on baseline setting.

Balancing Exploration and Exploitation

Pass@k Training not only improves exploration by maintaining higher entropy and diversity in response generation (Figures 6 and 7) but also exhibits strong generalization capabilities across various domains and tasks. This robustness makes it superior to typical noise-induced rewards or entropy regularization methods, which often lead to new trade-off challenges. Figure 4

Figure 4

Figure 4: Training progress of Noise Rewards and Entropy Regularization on baseline setting.

Figure 5

Figure 5

Figure 5: Diversity of Negative Responses on Maze Tasks.

Generalization and Robustness

The method demonstrates robust exploration capabilities across different task types and model architectures. Although increasing kk may reduce training efficiency, this can be mitigated with appropriate learning rate adjustments (Figure 6). Figure 6

Figure 6

Figure 6

Figure 6

Figure 6: Training progress of Pass@k Training under various k and learning rate (LR).

Practical Application and Transfer of Benefits

The benefits of Pass@k Training can be effectively transferred to enhance Pass@1 performance, as evidenced in various labyrinthine and mathematical reasoning tasks. This transferability underscores the potential of Pass@k Training to drive performance in LLMs beyond mere exploratory tasks.

Generalizing Pass@k Training via Implicit Reward Design

Analysis of Advantage Functions

The analysis reveals that the success of Pass@k Training is influenced by the position of the advantage function's maximum and its trend, suggesting that focusing on harder problems might improve optimization efficiency.

Implicit Reward Design

The study explores potential implicit reward designs, such as Exceeding Pass@k Training and a combination of Pass@1 and Pass@k Training, showing improved model performance. Adaptive strategies guided by policy entropy further enhance both exploration and exploitation capabilities (Figures 11-13). Figure 7

Figure 7

Figure 7: Training progress of Pass@k Training and Exceeding Pass@k Training on baseline setting.

Figure 8

Figure 8

Figure 8: Training progress of Pass@k Training and Combination Training on baseline setting.

Figure 9

Figure 9

Figure 9: Training progress of Pass@k Training and Adaptive Training on baseline setting.

Conclusion

Pass@k Training represents a significant advancement in balancing exploration and exploitation in RLVR for LRMs. By incorporating the Pass@k metric as a reward, the paper demonstrates its potential to enhance exploration without adversely affecting performance, paving the way for future research in RLVR training methodologies and implicit reward design.

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 10 tweets with 510 likes about this paper.