Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 148 tok/s
Gemini 2.5 Pro 47 tok/s Pro
GPT-5 Medium 26 tok/s Pro
GPT-5 High 28 tok/s Pro
GPT-4o 94 tok/s Pro
Kimi K2 214 tok/s Pro
GPT OSS 120B 429 tok/s Pro
Claude Sonnet 4.5 36 tok/s Pro
2000 character limit reached

One-Token Rollout: Guiding Supervised Fine-Tuning of LLMs with Policy Gradient (2509.26313v1)

Published 30 Sep 2025 in cs.CL

Abstract: Supervised fine-tuning (SFT) is the predominant method for adapting LLMs, yet it often struggles with generalization compared to reinforcement learning (RL). In this work, we posit that this performance disparity stems not just from the loss function, but from a more fundamental difference: SFT learns from a fixed, pre-collected dataset, whereas RL utilizes on-policy data sampled from the current policy. Building on this hypothesis, we introduce one-token rollout (OTR), a novel fine-tuning algorithm that guides SFT with the policy gradient method. OTR reframes the autoregressive learning process by treating each token generation as a single-step reinforcement learning trajectory. At each step, it performs a Monte Carlo ``rollout'' by sampling multiple candidate tokens from the current policy's distribution. The ground-truth token from the supervised data is then used to provide a reward signal to these samples. Guided by policy gradient, our algorithm repurposes static, off-policy supervised data into a dynamic, on-policy signal at the token level, capturing the generalization benefits of on-policy learning while bypassing the costly overhead of full sentence generation. Through extensive experiments on a diverse suite of challenging benchmarks spanning mathematical reasoning, code generation, and general domain reasoning, we demonstrate that OTR consistently outperforms standard SFT. Our findings establish OTR as a powerful and practical alternative for fine-tuning LLMs and provide compelling evidence that the on-policy nature of data is a critical driver of generalization, offering a promising new direction for fine-tuning LLMs.

Summary

  • The paper introduces One-Token Rollout, a token-level policy gradient method that bridges supervised fine-tuning and reinforcement learning to enhance LLM generalization.
  • It employs token-level sampling with a negative reward for incorrect tokens, which stabilizes training and reduces perplexity compared to conventional SFT.
  • Empirical evaluations across math, code, and reasoning benchmarks demonstrate improved performance, reduced catastrophic forgetting, and effective knowledge retention.

One-Token Rollout: A Policy Gradient Approach to Token-Level On-Policy Fine-Tuning of LLMs

Motivation and Theoretical Foundations

The paper introduces One-Token Rollout (OTR), a fine-tuning algorithm for LLMs that leverages policy gradient methods at the token level to address the generalization gap between supervised fine-tuning (SFT) and reinforcement learning (RL). SFT, while efficient and widely adopted, is fundamentally off-policy, relying on static expert demonstrations. In contrast, RL methods utilize on-policy data, which has been shown to yield superior generalization and better preservation of pre-trained knowledge, particularly in low-probability regions of the model's output space.

OTR is motivated by the hypothesis that the generalization disparity between SFT and RL is primarily due to the nature of the training data—static versus dynamic—rather than the loss function itself. By reframing each token generation as a single-step RL trajectory, OTR simulates on-policy learning within the SFT framework, thereby capturing the generalization benefits of RL without incurring the computational cost of full sequence rollouts.

Algorithmic Design and Implementation

OTR operationalizes policy gradient at the token level. For each token position tt in a training sequence, the algorithm samples KK candidate tokens from the current policy πθ\pi_\theta using a temperature-scaled softmax to encourage exploration. Each sampled token at,ja'_{t,j} is assigned a reward: $1$ if it matches the ground-truth token xtx_t, and β\beta (typically negative) otherwise. The per-token loss is then computed as:

LOTRt(θ)=1K[Ngtlogπθ(xtst)+βj:at,jxtlogπθ(at,jst)]\mathcal{L}_{\text{OTR}^t}(\theta) = -\frac{1}{K} \left[ N_{gt} \log \pi_\theta(x_t | s_t) + \beta \sum_{j: a'_{t,j} \neq x_t} \log \pi_\theta(a'_{t,j} | s_t) \right]

where NgtN_{gt} is the number of times the ground-truth token is sampled. This loss is averaged over all tokens in the sequence. The use of a negative β\beta penalizes the model for assigning high probability to incorrect tokens, functioning as a regularizer that improves generalization and stability. Figure 1

Figure 1: An illustration of the computational divergence between SFT and OTR, highlighting OTR's token-level on-policy sampling and reward assignment.

OTR is implemented efficiently: it requires only token-level sampling and reward computation, avoiding the need for full sequence generation or complex off-policy corrections. The algorithm is compatible with standard LLM training frameworks and can be integrated with minimal changes to the data pipeline.

Empirical Evaluation

Experiments are conducted on mathematical reasoning (OpenR1-Math-220k, NuminaMath-1.5), code generation (HumanEval+, MBPP+), and general reasoning (MMLU-Pro, SuperGPQA, BBEH) benchmarks using Qwen2.5 and Qwen3 model families (3B–8B parameters). OTR consistently outperforms SFT across all domains and model sizes, both in-domain and out-of-domain.

Key findings include:

  • In mathematical reasoning, OTR yields higher average scores than SFT for all tested models. It also demonstrates reduced catastrophic forgetting, with fewer instances of performance degradation below the base model.
  • In code and general reasoning tasks, OTR achieves superior average performance, indicating improved transferability and generalization.
  • Ablation studies on the reward hyperparameter β\beta show that negative values (e.g., β=0.1\beta = -0.1) provide optimal stability and performance. Assigning positive rewards to incorrect tokens leads to training instability and degraded model distributions. Figure 2

Figure 2

Figure 2: Effect of β\beta on GT token counts, demonstrating the impact of reward regularization on training stability and generalization.

OTR also converges to higher ground-truth token sampling rates than SFT, indicating lower perplexity and more effective utilization of training data.

OTR generalizes the dynamic fine-tuning (DFT) approach, which reweights the SFT loss for ground-truth tokens by their model probability. When β=0\beta = 0, OTR reduces to a Monte Carlo approximation of DFT. However, OTR's explicit penalization of incorrect tokens (β<0\beta < 0) yields superior empirical results, underscoring the importance of negative sampling in fine-tuning objectives.

OTR is distinct from prior RL-based fine-tuning methods (e.g., PPO, GRPO, DPO, GPG) in its computational efficiency and simplicity. It avoids the instability and complexity of full trajectory RL, while still capturing the essential benefits of on-policy learning.

Practical Implications and Future Directions

OTR provides a practical, scalable alternative to SFT for fine-tuning LLMs, with clear advantages in generalization, knowledge preservation, and computational efficiency. Its token-level design makes it suitable for large-scale training and easy integration into existing pipelines.

Potential future developments include:

  • Scaling OTR to larger models (e.g., 70B+ parameters) and broader datasets.
  • Extending the framework to multi-token rollouts and more sophisticated reward functions.
  • Adapting OTR to multimodal domains (e.g., vision-language tasks).
  • Investigating theoretical properties of token-level on-policy learning and its impact on model calibration and robustness.

Conclusion

One-Token Rollout bridges the gap between SFT and RL by simulating on-policy learning at the token level, yielding consistent improvements in generalization and knowledge retention across diverse LLM benchmarks. The approach demonstrates that dynamic, on-policy data is a critical driver of fine-tuning efficacy, and establishes OTR as a robust, efficient, and theoretically grounded alternative to standard SFT for LLM adaptation.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Youtube Logo Streamline Icon: https://streamlinehq.com