Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash 91 tok/s
Gemini 2.5 Pro 45 tok/s Pro
GPT-5 Medium 37 tok/s
GPT-5 High 35 tok/s Pro
GPT-4o 105 tok/s
GPT OSS 120B 463 tok/s Pro
Kimi K2 235 tok/s Pro
2000 character limit reached

MeanFlow-Based Generative Model

Updated 12 August 2025
  • MeanFlow-based models are generative frameworks that replace instantaneous velocity with an interval-averaged (mean) velocity, enabling direct mapping from noise to data.
  • The approach utilizes the MeanFlow Identity to relate average and instantaneous velocities, supporting one- or few-step generation with efficient training.
  • Empirical evaluations show state-of-the-art results in image synthesis, audio generation, and policy learning, highlighting practical benefits in speed and consistency.

A MeanFlow-based model refers to a class of generative and inference frameworks that replace the classic instantaneous velocity field used in flow matching with an interval-averaged (mean) velocity, thereby enabling direct, efficient mappings from noise to data. By formulating an explicit relationship—termed the MeanFlow Identity—between average and instantaneous velocities, these models support one- or few-step generation, rapid inference, and principled consistency, with robust empirical performance across vision, audio, policy learning, and other modalities (Geng et al., 19 May 2025, Sheng et al., 14 Jul 2025, Li et al., 8 Aug 2025, Guo et al., 22 Jul 2025, Cao et al., 9 Aug 2025).

1. Mathematical Foundations of MeanFlow

MeanFlow-based models are grounded in integral formulations of dynamical systems that parametrize flows not by their instantaneous velocities v(z,t)v(z, t), but by their average velocity u(zt,r,t)u(z_t, r, t) over an interval [r,t][r, t]:

u(zt,r,t)=1trrtv(zτ,τ)dτu(z_t, r, t) = \frac{1}{t - r} \int_r^t v(z_\tau, \tau) d\tau

A central theoretical result is the MeanFlow Identity, which relates the average velocity to the instantaneous velocity:

u(zt,r,t)=v(zt,t)(tr)ddtu(zt,r,t)u(z_t, r, t) = v(z_t, t) - (t - r)\frac{d}{dt}u(z_t, r, t)

Here, ddtu(zt,r,t)\frac{d}{dt}u(z_t, r, t) is decomposed into terms involving both zu\partial_{z} u and tu\partial_t u via the chain rule.

The MeanFlow model uses this identity as a training target: a neural network is trained to output uθ(z,r,t)u_\theta(z, r, t) such that, given a stochastic trajectory ztz_t (interpolated between data and noise), the network's output solves the identity above throughout training.

This approach allows the direct mapping from prior noise (e.g., z1N(0,I)z_1\sim \mathcal{N}(0, I)) to a data sample in a single inference step:

z0=z1uθ(z1,0,1)z_0 = z_1 - u_\theta(z_1, 0, 1)

2. Implementation Strategies and Loss Formulation

At training time, the following steps are performed:

  1. Draw pairs of times (r,t)(r, t) (with t>rt > r), a data sample xx, and latent noise ee.
  2. Compute the intermediate flow state z=(1t)x+tez = (1-t)\,x + t\,e.
  3. Calculate the conditional velocity vtv_t (for the "straight" flow schedule, usually vt=exv_t = e - x).
  4. To evaluate the right-hand side of the MeanFlow Identity, compute the Jacobian–vector product (JVP) to obtain derivatives of uθu_\theta with respect to zz and tt.
  5. The loss is defined as:

Loss(θ)=E[uθ(z,r,t)sg(vt(tr)(vtzuθ+tuθ))22]\operatorname{Loss}(\theta) = \mathbb{E}\left[ \| u_\theta(z, r, t) - \mathrm{sg}(v_t - (t - r)(v_t \cdot \partial_{z}u_\theta + \partial_t u_\theta)) \|_2^2 \right]

where sg()\mathrm{sg}(\cdot) denotes a stop-gradient operation to prevent higher-order derivatives during backpropagation.

No pretraining, distillation, or curriculum learning is required. For practical efficiency, time pairs (r,t)(r, t) are commonly sampled using a logit-normal or uniform distribution, and JVPs are computed using deep learning frameworks’ automatic differentiation (e.g., torch.func.jvp, jax.jvp).

At inference, one directly applies z0=z1uθ(z1,0,1)z_0 = z_1 - u_\theta(z_1, 0, 1), realizing one-step generation.

3. Empirical Performance and Applications

MeanFlow-based models have demonstrated state-of-the-art results in multiple domains:

  • Image Synthesis: Achieves Fréchet Inception Distance (FID) of 3.43 on ImageNet 256×256 with a single function evaluation (1-NFE), outperforming previous one-step diffusion and flow-based models (Geng et al., 19 May 2025).
  • Audio Synthesis: MeanAudio attains a real-time factor (RTF) of 0.013—enabling a 100×100\times inference speedup over previous diffusion-based systems, while preserving synthesis quality (Li et al., 8 Aug 2025).
  • Robotic Policy Learning: MP1, which employs the MeanFlow paradigm, obtains superior task success (e.g., 10.2% better than DP3 on the Adroit and Meta-World benchmarks) and achieves 19×\times faster inference relative to iterative diffusion policy models (Sheng et al., 14 Jul 2025).
  • Reinforcement Learning: MeanFlow policy parametrizations in Flow Policy Mirror Descent yield comparable MuJoCo benchmark performance to diffusion policies while reducing the number of inference function evaluations by several orders of magnitude (Chen et al., 31 Jul 2025).
  • Recommender Systems: FMRec leverages a similar flow matching structure for deterministic and efficient sequential recommendations (Liu et al., 22 May 2025).
  • Speech Synthesis: SplitMeanFlow, an algebraic generalization, achieves 20×20\times speedup in real-world text-to-speech applications (Guo et al., 22 Jul 2025).

MeanFlow enables one-step or few-step sampling with little or no degradation in generative quality compared to iterative models.

4. Extensions and Algorithmic Enhancements

Several advancements generalize or refine the MeanFlow approach:

  • Interval Splitting Consistency (SplitMeanFlow): Replaces derivative-based objectives with a purely algebraic constraint reflecting integral additivity. The Interval Splitting Consistency is formulated as:

(tr)u(zt,r,t)=(sr)u(zs,r,s)+(ts)u(zt,s,t)(t-r)u(z_t, r, t) = (s-r)u(z_s, r, s) + (t-s)u(z_t, s, t)

for any r<s<tr < s < t, and is leveraged for more stable and efficient training without requiring Jacobian computations (Guo et al., 22 Jul 2025).

  • High-Order MeanFlow: Second-Order MeanFlow further incorporates average acceleration, uses a generalized additive identity for average acceleration, and demonstrates that the resulting sampling algorithm resides in TC0\mathsf{TC}^0 (constant-depth threshold circuits), optimizing both expressivity and hardware efficiency (Cao et al., 9 Aug 2025).
  • Classifier-Free Guidance (CFG): MeanAudio and MP1 tightly integrate CFG directly into their training objectives rather than as a post-hoc sampling modification, allowing for controllable generations without increasing inference cost (Li et al., 8 Aug 2025, Sheng et al., 14 Jul 2025).
  • Curriculum and Mix-up Strategies: MeanAudio introduces an instantaneous-to-mean curriculum, blending standard (instantaneous) and mean flow matching during training for increased stability and convergence rate (Li et al., 8 Aug 2025).
  • Dispersive Loss: MP1 adds a loss that repels encoded representations for different states, aiding generalization in low-data regimes (Sheng et al., 14 Jul 2025).
  • Deterministic Reverse Sampling: FMRec uses a straight-flow ODE with an Euler solver, which is exact for linear trajectories, further reducing sampling noise and inference cost (Liu et al., 22 May 2025).

5. Theoretical Implications and Computational Guarantees

MeanFlow-based models and their algebraic generalizations provide several guarantees and theoretical properties:

  • Consistency: Satisfaction of the MeanFlow identity (or its algebraic analogs) ensures that the generated trajectories are self-consistent, either in the differential (MeanFlow) or algebraic (SplitMeanFlow) sense.
  • Sampling Efficiency: By summarizing the entire flow as an average (rather than integrating small steps), one-step or few-step sampling is possible without introducing large discretization errors. In policy learning, the discretization error in one-step sampling is controlled by the variance of the target distribution and vanishes as the policy approaches determinism (Chen et al., 31 Jul 2025).
  • Expressivity: The circuit complexity of second-order MeanFlow sampling (via transformer networks) remains in constant-depth, polynomial-size threshold circuits, ensuring practical scalability even with richer dynamical representations (Cao et al., 9 Aug 2025).
  • Hardware Compatibility: The move towards algebraic objectives (as in SplitMeanFlow) and fast approximate attention mechanisms enables improved scalability on both conventional and specialized accelerator hardware (Guo et al., 22 Jul 2025, Cao et al., 9 Aug 2025).

6. Broader Impact and Future Directions

The MeanFlow-based paradigm has contributed to closing the gap between one-step and classic multi-step generative models, particularly in tasks requiring real-time inference and efficient computation (e.g., robotics, text-to-audio, policy learning).

Future directions include:

  • Further exploration of high-order (e.g., second-order) MeanFlow objectives, leveraging curvature and higher-order dynamics for increased expressivity without trading off sampling efficiency (Cao et al., 9 Aug 2025).
  • Generalized integral consistency principles (Interval Splitting Consistency) that bypass the need for Jacobian computations, making deployment more robust and accessible (Guo et al., 22 Jul 2025).
  • Cross-domain applications in simulation-based modeling, data assimilation in the physical sciences, and reinforcement learning exploitation–exploration dynamics.

By rigorously connecting integral consistency, hardware efficiency, and generative expressivity, MeanFlow-based models provide a theoretically principled and practically powerful foundation for modern high-speed generative modeling and inference.