Papers
Topics
Authors
Recent
Search
2000 character limit reached

Stochastic Interpolants

Updated 15 July 2025
  • Stochastic interpolants are mathematical constructs that define a continuous stochastic process accurately linking two probability distributions via deterministic and diffusion dynamics.
  • They unify normalizing flows and score-based diffusions using explicit formulations for drift and score functions optimized with quadratic loss objectives.
  • They are pivotal in generative modeling, high-dimensional sampling, and forecasting, with extensions to multimarginal, Riemannian, and latent variable settings.

Stochastic interpolants are mathematical constructs and algorithmic frameworks designed to generate a continuous or stochastic process that exactly bridges two prescribed probability distributions in a finite time interval. Originally developed to unify and extend normalizing flows and score-based diffusion models, stochastic interpolants have emerged as a central paradigm in generative modeling, scientific computing, probabilistic forecasting, and high-dimensional sampling. Their flexibility allows the realization of both deterministic transport and stochastic diffusion-based transformations, and modern research has generalized them to multimarginal, Riemannian, data-dependent, and latent-variable settings.

1. General Framework of Stochastic Interpolants

At the core, a stochastic interpolant is a time-indexed family of random variables {xt}t[0,1]\{x_t\}_{t\in[0,1]} constructed so that x0x_0 follows an initial (“base”) distribution ρ0\rho_0, x1x_1 follows a target distribution ρ1\rho_1, and the path between these endpoints defines an explicit, typically smooth, interpolation in probability space. The canonical form is

xt=I(t,x0,x1)+γ(t)zx_t = I(t, x_0, x_1) + \gamma(t) z

with x0ρ0x_0 \sim \rho_0, x1ρ1x_1 \sim \rho_1, zN(0,I)z \sim \mathcal{N}(0, I), and I(0,x0,x1)=x0I(0, x_0, x_1) = x_0, x0x_00, x0x_01 (Albergo et al., 2023). Various choices of x0x_02 (e.g., linear, trigonometric) and x0x_03 (deterministic or vanishing at endpoints) allow the specification of the “type” of interpolant: from strictly deterministic (flow-based) to fully stochastic (diffusion-based).

The time-evolving marginal densities x0x_04 of x0x_05 satisfy a first-order continuity equation

x0x_06

where x0x_07 is the velocity field obtained by conditional expectation of the time derivative given x0x_08: x0x_09 Including stochasticity introduces a diffusion term so that ρ0\rho_00 satisfies a (forward or backward) Fokker–Planck equation (Albergo et al., 2023).

These interpolants generalize both deterministic optimal transport (e.g., normalizing flows) and noise-driven generative models (e.g., score-based diffusions) (Albergo et al., 2022). The unifying mathematical structure enables flexible, model-agnostic specification of sampling, transform, and generative operations.

2. Mathematical Formulation and Learning

The functional form of ρ0\rho_01, ρ0\rho_02, and the endogenous (or exogenous) noise term ρ0\rho_03 determines the character of the interpolant:

  • Linear interpolant: ρ0\rho_04, with ρ0\rho_05 possibly ρ0\rho_06 (Albergo et al., 2023), recovers ordinary flows when ρ0\rho_07 and classical diffusions when ρ0\rho_08.
  • Trigonometric or other smooth interpolants: ρ0\rho_09 (Albergo et al., 2022).

The drift x1x_10 and the score function x1x_11 are key quantities for constructing transport and sampling algorithms. These are framed as minimizers of explicit quadratic objective functions: x1x_12 and similarly for the score (Albergo et al., 2023, Albergo et al., 2022). These objectives enable training via regression or empirical risk minimization, sidestepping deep simulation-based losses.

Sampling is conducted by integrating the learned drift or joint drift-score fields using forward or reverse ODEs or SDEs:

  • Deterministic (ODE/“probability flow”): x1x_13.
  • Stochastic (SDE): x1x_14, where x1x_15.

This approach covers both exact sampling (when x1x_16 and x1x_17 are learned perfectly) and approximate settings, with theoretical bounds on bias and discretization error (Liu et al., 13 Feb 2025, Wu et al., 22 Apr 2025).

3. Generalizations and Extensions

The stochastic interpolant framework has been extended substantially:

  • Multimarginal Interpolants: By parameterizing the interpolation with a simplex coordinate x1x_18 rather than scalar x1x_19, the method bridges more than two densities simultaneously, yielding a process ρ1\rho_10 over the simplex ρ1\rho_11, enabling applications like all-to-all style transfer and algorithmic fairness (Albergo et al., 2023).
  • Data-dependent Couplings: By learning or specifying couplings ρ1\rho_12 beyond product form, the interpolant can define conditional generative models, improving performance in tasks such as super-resolution and in-painting (Albergo et al., 2023).
  • Riemannian Manifolds: For settings like ρ1\rho_13 (sphere), SO(3) (rotations), and general Riemannian manifolds, geodesic-based interpolants ρ1\rho_14 are used, and the marginal flows are governed by transport equations on the manifold (Wu et al., 22 Apr 2025).
  • Latent Variable Models: Latent Stochastic Interpolants (LSI) define interpolations in learned latent spaces, enabling joint training of encoder, decoder, and SI modules via a continuous-time Evidence Lower Bound (ELBO) (Singh et al., 2 Jun 2025).

4. Algorithms and Practical Implementations

Typical implementation of stochastic interpolants involves learning neural parameterizations of ρ1\rho_15 and ρ1\rho_16 via regression on samples ρ1\rho_17, with losses computed as expectations in ρ1\rho_18 (for instance, equispaced or importance sampled in ρ1\rho_19) (Albergo et al., 2023, Albergo et al., 2022).

Sampling Strategies:

  • Forward ODE/SDE integration: Numerically solve xt=I(t,x0,x1)+γ(t)zx_t = I(t, x_0, x_1) + \gamma(t) z0 or its stochastic version.
  • Discrete-time approximation: Euler–Maruyama schemes are analyzed with finite-time KL error bounds, with scheduling strategies (e.g., exponentially decaying timestep sizes) designed to control discretization error due to stiff interpolants and small latent noise scales (Liu et al., 13 Feb 2025).
  • Conditional/Multimarginal models: Learn conditional velocities xt=I(t,x0,x1)+γ(t)zx_t = I(t, x_0, x_1) + \gamma(t) z1 for each marginal and interpolate as xt=I(t,x0,x1)+γ(t)zx_t = I(t, x_0, x_1) + \gamma(t) z2 (Albergo et al., 2023).
  • Energy-consistent (Physics-aware) Interpolants: Parameterize xt=I(t,x0,x1)+γ(t)zx_t = I(t, x_0, x_1) + \gamma(t) z3, xt=I(t,x0,x1)+γ(t)zx_t = I(t, x_0, x_1) + \gamma(t) z4 in (possibly Fourier) bases and optimize to preserve physical invariants (such as kinetic energy), crucial for long fluid dynamics rollouts (Mücke et al., 8 Apr 2025).

Theoretical Analysis:

  • Contractivity: For Gaussian to log-concave targets, the SI flow map is Lipschitz with constants matching those from Caffarelli’s theorem for optimal transport. This ensures stability and robustness for sampling and estimation (Daniels, 14 Apr 2025).
  • Generative Bias on Manifolds: KL-divergence between generated and target marginal laws depends on mismatches in learned velocity and score fields, with explicit bounds given in terms of time-integrated inner products (Wu et al., 22 Apr 2025).

Extension to High Dimensions:

Machine learning approaches—including neural ODEs, denoising networks, or neural FBSDE solvers—allow tractable implementation even for xt=I(t,x0,x1)+γ(t)zx_t = I(t, x_0, x_1) + \gamma(t) z5 (George et al., 1 Feb 2025).

5. Applications in Generative Modeling and Scientific Computing

Stochastic interpolants have enabled advances in numerous domains:

  • Unifying Flows and Diffusions: The framework is the first to provably and practically connect flow-based and diffusion-based models, allowing exact finite-time mappings between arbitrary base and target densities (Albergo et al., 2023).
  • Multimodal and Conditional Generation: Can be applied to multimarginal problems (all-to-all translation, fair generation), and to conditional tasks (super-resolution, in-painting) via data-coupled interpolants (Albergo et al., 2023, Albergo et al., 2023).
  • Material and Molecular Generation: Used as the generative core in open-ended material discovery (Open Materials Generation) and Boltzmann sampling (e.g., BoltzNCE), with state-of-the-art accuracy and efficiency (Hoellmer et al., 4 Feb 2025, Aggarwal et al., 1 Jul 2025).
  • Protein and Fluid Dynamics Simulations: SI-based models (with SO(3)-equivariance or energy-consistency) enable accelerated molecular dynamics and long-horizon, stable fluid simulations, outperforming classical and contemporary deep learning models (Costa et al., 2024, Mücke et al., 8 Apr 2025).
  • Time Series and Forecasting: Stochastic interpolants are combined with recurrence and SDEs for efficient probabilistic forecasting of multivariate time series and high-dimensional physical systems, including Föllmer (optimal transport) sampling for conditional distributions (Chen et al., 2024, Chen et al., 2024).

6. Theoretical and Practical Implications

The interpolant framework offers explicit, simulation-free (quadratic loss) training objectives and has clarified the relationship between transport, diffusion, and score functions. The contractivity and monotonicity properties derived in recent work have enabled precise control of sampling error, stability, and regularity, matching optimal transport and functional inequality bounds (Daniels, 14 Apr 2025, Liu et al., 13 Feb 2025).

Recent extensions to Riemannian manifolds and latent-variable models allow for learning on complex geometric domains and for matching in joint latent-observation spaces, leveraging advanced sampling schemes such as embedding-SDEs and flexible ELBOs (Wu et al., 22 Apr 2025, Singh et al., 2 Jun 2025).

Limitations include the potential for numerical instability near endpoints (when the latent noise scale vanishes), increased computational costs in score estimation (for high-dimensional targets), and open questions regarding scalability to very large molecular or materials systems using current neural architectures (Aggarwal et al., 1 Jul 2025). Ongoing research addresses adaptive time stepping, coupling structure, further physical invariants, and broader classes of governing SDEs.

7. Outlook and Future Directions

The stochastic interpolant paradigm is anticipated to underpin the next generation of generative models for scientific, geometric, and structured probabilistic data. Promising directions include enhanced coupling for inverse problems, hybridization with transformer and sequence models for temporal prediction, augmentation with automatic symmetry and constraint incorporation, and deeper integration with control-theoretic methods for stochastic process design (Costa et al., 2024, Hoellmer et al., 4 Feb 2025, Wu et al., 22 Apr 2025).

The flexibility and theoretical foundation provided by stochastic interpolants—expressed in mathematical formulations for velocity, score, and transport fields; empirical risk minimization algorithms; and contractivity theorems for sampling—form a robust basis for future unified approaches to high-fidelity generative modeling across statistical, physical, and data-driven settings.

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Stochastic Interpolants.