Reparameterization Flow Policy Optimization
- RFO is a reinforcement learning method that trains continuous normalizing flow policies using reparameterization gradients, bypassing intractable likelihood evaluations.
- It employs Conditional Flow Matching regularization to ensure stability through past-data consistency and enhanced exploration via uniform-target adjustments.
- Empirical results show RFO achieves state-of-the-art performance in complex control tasks, outperforming prior baselines with improved sample efficiency and robustness.
Reparameterization Flow Policy Optimization (RFO) is a reinforcement learning (RL) methodology for training highly expressive, continuous normalizing flow (CNF) policies, leveraging the reparameterization (pathwise derivative) gradient to enable stable and sample-efficient learning. RFO unifies flow-based policy parameterizations with differentiable (often model-based) RL by allowing gradients to be backpropagated directly through the flow-generator ODEs and, if present, the environment dynamics, yielding policy optimization without intractable likelihood evaluations. The RFO framework, introduced in (Zhong et al., 3 Feb 2026), integrates regularization terms for stability and exploration, and in practical instantiations, supports both on-policy and off-policy algorithm designs. RFO has demonstrated state-of-the-art performance in diverse continuous control and manipulation tasks.
1. Foundations: Reparameterization Policy Gradients and Flow Policies
The reparameterization policy gradient (RPG) framework in model-based RL exploits differentiable simulators to compute control gradients via the pathwise (reparameterization) method rather than classic likelihood-ratio approaches. The policy is represented as
with expected return objective
Unlike standard Gaussian policies, flow-based policies use a state-conditioned CNF to transform base noise to actions via the ODE
with the final action . This framework allows flexible, potentially multimodal action sampling, but naive application with RPG leads to instability and poor exploration, necessitating specialized regularization schemes (Zhong et al., 3 Feb 2026).
2. Flow Policy ODE Parameterizations and Training Without Log-Likelihoods
A flow policy parameterizes the stochastic mapping from base noise to agent actions as an invertible, differentiable time-dependent ODE flow: The action-conditional density is formally given by the change-of-variables formula, but RFO sidesteps the computation of the inverse flow and associated Jacobian determinants: Instead, RFO optimizes solely by direct backpropagation through the ODE-solver and, as applicable, the system dynamics, avoiding any log-likelihood evaluations (Zhong et al., 3 Feb 2026).
Practical implementations adopt numerical integration (Euler or higher-order methods, typically steps), and rely on deep neural approximations for the velocity field .
3. Stability and Exploration: Conditional Flow Matching Regularization
Training instability arises from the invertibility of the flow ODE and the possibility of subsequent updates breaking action trajectories previously visited—making the policy unable to re-sample those actions. RFO introduces two Conditional Flow Matching (CFM) regularization losses:
- Past-Data CFM Regularization (Stability):
This term ensures the vector field remains consistent with recently visited action trajectories.
- Uniform-Target CFM Regularization (Exploration):
This term encourages the policy to cover the entirety of the action space, improving exploration.
The overall loss combines the (negative) short-horizon return with the above regularizers. Empirical ablations show that both are essential for robust optimization; omitting either degrades performance to earlier RPG baselines (Zhong et al., 3 Feb 2026).
4. Algorithmic Structure and Action Chunking Variant
The canonical RFO training loop follows the Short-Horizon Actor-Critic (SHAC) style:
- Collect short-horizon rollout trajectories under the current policy.
- Augment recent-action and rollout buffers for use in CFM regularization.
- Compute the reparameterized gradient of the short-horizon return proxy by BPTT through both the flow and simulator.
- Apply and regularization gradients.
- Update policy parameters and critic parameters .
The action-chunking variant extends the flow policy to emit action sequences of length at each step,
which are executed in order before the next observation. The optimization is carried out on blocked segments, with the flow ODE and its regularizers extended to the higher-dimensional chunked action space (Zhong et al., 3 Feb 2026).
5. Empirical Results and Comparative Evaluation
RFO has been validated on a wide spectrum of RL benchmarks with differentiable physics (DFlex, Rewarped):
- Locomotion: Ant (), ANYmal (), Soft Jumper (, visual).
- Manipulation: Hand Reorient, Rolling Pin (pixel/state), Hand Flip, Transport.
The benchmarks include both state and pixel-based tasks and both rigid and soft-body dynamics.
In all reported environments, RFO achieves competitive or strictly superior final returns relative to SHAC, SAPO, FlowRL, DrAC, and other diffusion-based or flow-matching baselines. Notably, in the high-dimensional Soft Jumper, RFO attains nearly the return of the best previous method. The following table presents mean normalized performance (SHAC=1.0):
| Task | SHAC | SAPO | RFO (ours) |
|---|---|---|---|
| Soft Jumper | 1.00 | 1.39 | 2.63 |
| Ant | 1.00 | 1.55 | 1.81 |
| Hand Reorient | 1.00 | 1.22 | 1.48 |
| ANYmal | 1.00 | 0.97 | 1.07 |
| Transport | 1.00 | 1.05 | 1.87 |
| Rolling Pin | 1.00 | 1.03 | 1.06 |
| Hand Flip | 1.00 | 1.27 | 1.26 |
Additionally, ablation studies confirm the necessity of both CFM regularization components for robust, stable learning (Zhong et al., 3 Feb 2026).
6. Theoretical Insights and Limitations
RFO uses the pathwise derivative throughout the combined flow and simulator graph, which yields low-variance gradient estimates characteristic of reparameterization but without requiring log-likelihood or density estimation. Unlike surrogate likelihood-ratio methods, RFO relies exclusively on differentiability of both the flow policy and environment dynamics.
The regularization terms are formulated as flow-matching objectives that guarantee stability (by maintaining reachability of previously visited actions) and promote global exploration (by injecting probability mass throughout the action space). RFO does not require explicit normalization constants or Jacobian determinants.
The method presupposes access to a fully differentiable simulator or a learned differentiable transition model. Backpropagation through long horizons may be computationally expensive but is partially mitigated via short-horizon proxy objectives. Action-chunked variants are supported but present greater optimization difficulty (Zhong et al., 3 Feb 2026).
7. Connections, Related Work, and Future Directions
RFO generalizes prior reparameterization-based policy gradient algorithms by supporting expressive, multimodal CNF policies. Unlike PolicyFlow (Yang et al., 1 Feb 2026), which targets on-policy PPO via a reparameterized importance-ratio surrogate combined with a Brownian-motivated entropy regularizer, RFO is fundamentally pathwise and sidesteps likelihood-ratio computations and their typical numerical instability. Off-policy RFO instantiations, as in SAC Flow (Zhang et al., 30 Sep 2025), further exploit velocity reparameterizations (Flow-G, Flow-T) to overcome gradient pathologies arising from the underlying residual RNN structure of flow rollouts.
Research directions highlighted in the literature include offline-to-online pretraining for flow policies, hybridizing with diffusion-based policies, and integrating with non-ODE generative models for richer policy classes (Zhong et al., 3 Feb 2026, Zhang et al., 30 Sep 2025).
RFO establishes a principled and empirically validated framework for policy optimization with high-capacity flows, providing stable, efficient, and expressive solutions for contemporary RL control challenges.