Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 154 tok/s
Gemini 2.5 Pro 40 tok/s Pro
GPT-5 Medium 25 tok/s Pro
GPT-5 High 21 tok/s Pro
GPT-4o 93 tok/s Pro
Kimi K2 170 tok/s Pro
GPT OSS 120B 411 tok/s Pro
Claude Sonnet 4.5 36 tok/s Pro
2000 character limit reached

Group-Relative Policy Optimisation (GRPO)

Updated 20 October 2025
  • GRPO is a reinforcement learning framework that compares group-wise candidate actions to compute normalized advantages for robust policy updates.
  • It replaces single-sample reward baselines with relative, group-normalized advantages to improve stability and sample efficiency in LLM alignment and control tasks.
  • GRPO’s variants—including Hybrid, multi-objective, and continuous control adaptations—demonstrate practical strengths in managing complex reward structures and multi-agent environments.

Group-Relative Policy Optimisation (GRPO) is a reinforcement learning framework that replaces single-sample, absolute reward baselines with group-wise, relative comparison of candidate actions or responses. Instead of evaluating and updating a policy based on the reward signal from a single outcome, GRPO generates a set (group) of candidate outputs for a given state/prompt, computes a group-normalized advantage for each sample by comparing its reward to the group mean (and variance), and uses these relative advantages to drive policy updates. This approach balances the trade-off between stability (as in value-baseline methods) and expressiveness (as in empirical return-based RL), and has found especially prominent usage in LLM alignment, sequence modeling, continuous control, multi-objective settings, and structured multi-agent environments.

1. Core Principles and Mathematical Foundations

GRPO fundamentally redefines policy advantage estimation. In classical RL algorithms—such as Proximal Policy Optimization (PPO)—the advantage is computed using a value function baseline, for example,

AT=r(sT,aT)+γV(sT+1)V(sT)A_T = r(s_T, a_T) + \gamma V(s_{T+1}) - V(s_T)

where V()V(\cdot) is a learned value function. In contrast, GRPO estimates advantage empirically using a group of GG sampled actions or responses {o1,...,oG}\{o_1, ..., o_G\} for a given state, prompt, or context qq: Ai=riμGσGA_i = \frac{r_i - \mu_G}{\sigma_G} where μG=1Gj=1Grj\mu_G = \frac{1}{G} \sum_{j=1}^G r_j and σG\sigma_G is the group standard deviation.

The typical policy update objective in GRPO is: JGRPO(θ)=Eq,{oi}πold[1Gi=1Gmin(πθ(oiq)πold(oiq)Ai,clip(πθ(oiq)πold(oiq),1ϵ,1+ϵ)Ai)βDKL(πθπref)]J_{GRPO}(\theta) = \mathbb{E}_{q, \{ o_i \} \sim \pi_{old}} \left[ \frac{1}{G} \sum_{i=1}^G \min\left( \frac{\pi_\theta(o_i|q)}{\pi_{old}(o_i|q)} A_i, \mathrm{clip}(\frac{\pi_\theta(o_i|q)}{\pi_{old}(o_i|q)}, 1-\epsilon, 1+\epsilon) A_i \right) - \beta D_{KL}(\pi_\theta || \pi_{ref}) \right] This critic-free, group-normalized baseline provides robustness against scale/bias errors in reward models and eliminates the requirement of training a difficult or expensive value baseline.

2. Application Modalities and Variants

2.1. LLM Alignment and RLHF

In the context of LLMs, GRPO has become a leading alternative to PPO or Direct Preference Optimization (DPO) for reinforcement learning from human feedback (RLHF). By comparing multiple sampled responses for each prompt and assigning relative advantages, GRPO can optimize for reward signals from binary (verifiable), learned, or multi-objective reward models—allowing efficient, critic-free alignment (Mroueh, 9 Mar 2025, Li et al., 26 Mar 2025, Wu et al., 1 Oct 2025). This approach is empirically shown to stabilize training and improve sample efficiency in LLM fine-tuning versus baseline PPO.

2.2. Multi-Sample and Hybrid Methods

Hybrid GRPO extends PPO by combining group-level empirical reward averaging with value function-based bootstrapping. The advantage in Hybrid GRPO is given by: AT=1Nt=1N[f(r(sT,at))+V(st+1)]V(sT)A_T = \frac{1}{N} \sum_{t=1}^N [f(r(s_T, a_t)) + V(s_{t+1})] - V(s_T) where f(r)f(r) is a reward transformation (e.g., tanh\tanh) for normalization (Sane, 30 Jan 2025). Hybrid formulations achieve improved convergence, stability, and sample efficiency.

2.3. Continuous Control and Robotics

GRPO has been extended to continuous control by adapting group-normalization to trajectory-level policy clusters, state-aware advantage estimation, and regularization (temporal smoothness, inter-group diversity). The continuous extension employs clustering of full trajectories and state distributions (using k-means and DBSCAN) for robust advantage computation (Khanda et al., 25 Jul 2025).

2.4. Multi-Objective and Reward Hacking Mitigation

GRPO is vulnerable to “reward hacking” in multi-objective settings: reward functions with higher variance can dominate the group-normalized advantage. MO-GRPO resolves this by separately normalizing each reward function before aggregation so that all objectives contribute evenly: Ag(MO)=i=1KRi(q,og)μiσiA^{(MO)}_g = \sum_{i=1}^K \frac{R_i(q, o_g) - \mu_i}{\sigma_i} where μi,σi\mu_i, \sigma_i denote mean and std of RiR_i over the group (Ichihara et al., 26 Sep 2025).

2.5. Connection to Contrastive Learning and Minimal Group Size

GRPO’s group-normalized advantage can be reframed as a contrastive learning loss, where positive and negative samples are compared within a group. With binary (verifiable) rewards and group size G=2G=2, 2-GRPO recovers an unbiased gradient (modulo scaling) identical to DPO’s contrastive gradient, but with just two samples per prompt: A+=1,A=1A^+ = 1, \quad A^- = -1 Theoretical and empirical evidence suggests that 2-GRPO matches 16-GRPO’s performance at a fraction of the compute cost (Wu et al., 1 Oct 2025).

3. Methodological Extensions and Regularization

Enhancements to GRPO extend its application and address key challenges:

  • Entropy Regularization: Additive entropy terms in the policy objective to encourage exploration and robustness, particularly in dynamic or sparse environments (Sane, 30 Jan 2025).
  • Hierarchical Sub-Sampling: Incorporation of multi-step or hierarchical returns to account for longer-term dependencies in control or language generation (Sane, 30 Jan 2025).
  • Noise-Aware Reweighting: Stable GRPO (S-GRPO) introduces advantage reweighting calibrated to a symmetric noise model, mitigating the “think-answer mismatch” seen in reasoning tasks under noisy binary rewards (Shen et al., 8 Aug 2025).
  • Conflict-Aware Gradient Filtering: Group Trajectory Policy Optimization (GTPO) identifies tokens whose gradient updates would conflict due to positive/negative rewards and either masks or amplifies updates; entropy filtering further guards against policy collapse (Simoni et al., 5 Aug 2025).
  • Token Preference Learning: λ\lambda-GRPO introduces a learnable adaptability parameter to control token-level weighting in loss aggregation, correcting length bias inherent in vanilla GRPO (Wang et al., 8 Oct 2025).
  • Training-Free GRPO: Bypasses parameter updates entirely by iteratively building an experiential knowledge token prior in context, updating it via “semantic advantage” meta-prompting and library distillation (Cai et al., 9 Oct 2025).

4. Application Domains and Empirical Results

4.1. LLM Alignment and Reasoning

GRPO and Hybrid GRPO approaches have demonstrated improved reasoning capabilities and stability for LLMs. Experiments show that multi-sample and group-relative advantage computation enhances sample efficiency, reduces gradient variance, and speeds up convergence compared to PPO and single-sample RLHF baselines (Sane, 30 Jan 2025, Mroueh, 9 Mar 2025, Ding et al., 5 Jun 2025, Pang et al., 4 Aug 2025).

4.2. Vision and Speech

GRPO has been adapted for image captioning (Liang, 3 Mar 2025), visual generative modeling (including diffusion and VAR models) (Xue et al., 12 May 2025, Gallici et al., 29 May 2025), and both ASR (Shivakumar et al., 2 Sep 2025) and TTS (Liu et al., 23 Sep 2025) by formulating the group-wise reward normalization and advantage estimation in domain-appropriate architectures. In visual domains, group-normalized RLHF outperforms deterministic or supervised baselines (e.g., up to +181% on VideoAlign scores for video synthesis (Xue et al., 12 May 2025)).

4.3. Hyperparameter Optimization

GRPO is integrated with Transformer-based models in GRPOformer for hyperparameter optimization, where group-wise comparisons drive efficient trajectory updates; regularization against policy churn ensures robust performance improvements compared to purely trajectory-history Transformer baselines (Guo et al., 21 Sep 2025).

4.4. Multi-Agent and Structured Environments

In multi-agent scenarios such as spatial public goods games, GRPO-GCC adds a global cooperation constraint to the group-normalized policy updates, shaping collective incentives and preventing collapse to trivial equilibria (Yang et al., 7 Oct 2025).

Domain GRPO Variant Key Metrics Improved
LLMs (Hybrid) GRPO, 2-GRPO RLHF accuracy, sample efficiency
Visual Generation DanceGRPO HPS-v2.1, CLIP, VideoAlign scores
Continuous Control Continuous GRPO Return, stability (locomotion)
Speech Recognition GRPO WER, hallucinations
Multi-Objective RL MO-GRPO Balance/multi-metric optimization
Hyperparam Opt GRPOformer Beat the Random, Mean Perf.
Multi-Agent Societies GRPO-GCC Cooperation rate, variance

5. Limitations, Trade-Offs, and Open Research Problems

While GRPO offers robustness, stability, and critic-free implementation, several limitations and open questions are noted:

  • Variance and Sample Efficiency: GRPO may require increased group sizes for variance reduction, but theoretical insights show only two rollouts may suffice in contrastive (binary) reward settings (Wu et al., 1 Oct 2025).
  • Reward Hacking: Naïve aggregation in multi-objective reward structures leads to reward hacking; appropriate normalization or principled aggregation (MO-GRPO) is necessary (Ichihara et al., 26 Sep 2025).
  • Length Bias in Sequence Tasks: Uniform advantage across response tokens biases toward verbosity; λ\lambda-GRPO addresses this but the optimality of different token weighting strategies remains an open question (Wang et al., 8 Oct 2025).
  • Stability Under Noise: S-GRPO's noise-aware weighting effectively stabilizes training under high reward noise; how to generalize this to more complex, step-level or process-level supervision is open (Shen et al., 8 Aug 2025).
  • Scalability and Efficiency: Despite algorithmic improvements, hardware and generation bottlenecks (especially for large groups or long sequences in GRPO) remain a challenge; speculative decoding and concurrency-aware accelerations are being developed (Zhang et al., 26 Sep 2025).

6. Future Directions

Research frontiers for GRPO include:

  • Adaptive and Dynamic Clustering: In continuous and multi-task RL, adaptive clustering of policies, returns, or state distributions may further improve group normalization.
  • Process Supervision and Self-Correction: Multi-layer or process-decomposed GRPO (as in MGRPO) enables self-correction without dense causal supervision; recursive or deeper correction architectures are anticipated (Ding et al., 5 Jun 2025).
  • Contrastive Learning and Minimal Rollout Regimes: Exploiting the equivalence between contrastive losses and group-relative estimation may enable even more resource-efficient policy optimization.
  • Contextual and Training-Free Policy Adaptation: Training-free GRPO approaches suggest that carefully learned context priors may substitute for costly weight updates, especially where data and compute are limited (Cai et al., 9 Oct 2025).
  • Real-World Deployment and Multi-Agent Systems: Application to decentralized, large-scale, and high-dimensional settings demands further advances in group construction, normalization, and stability.

GRPO has emerged as a robust and adaptable RL framework, offering theoretically grounded, sample-efficient, and practically effective mechanisms for policy optimization in a wide range of discrete, continuous, and structured problems. Its innovations in group-normalized advantage estimation, critic-free updates, and flexibility in reward modeling have established new baselines for LLM alignment, vision and speech generation, multi-objective optimization, and beyond. Continued open questions revolve around further efficiency, stability, reward compositionality, and process-level adaptation.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (20)
Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Group-Relative Policy Optimisation (GRPO).

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube