Latent Stochastic Interpolants
- LSI is a framework that extends stochastic interpolants into latent space, enabling efficient generative modeling with continuous-time transport between arbitrary priors and data-driven posteriors.
- It jointly optimizes the encoder, decoder, and latent dynamics using a unified SDE/ODE formulation, reducing computational cost and enhancing training efficiency.
- LSI achieves competitive performance in generative tasks by lowering sampling FLOPs and supporting flexible latent representations, as validated by empirical and numerical analyses.
Latent Stochastic Interpolants (LSI) are a theoretical and algorithmic extension of the stochastic interpolant (SI) framework, transferring SI’s continuous-time transport between probability measures from data or observation space into a learned, low-dimensional latent space. This facilitates highly flexible generative modeling, supports arbitrary latent priors, and enables efficient sampling and training schemes using a unified SDE/ODE formalism. LSI jointly optimizes the encoder, decoder, and latent-path dynamics, allowing end-to-end learning of both latent representations and the generative process bridging an arbitrary prior to the data-driven aggregated posterior. This section details the mathematical formulation, algorithmic structure, sample complexity, numerical schemes, and salient properties of the LSI framework, referencing definitions and empirical results as established across the literature, principally in (Singh et al., 2 Jun 2025, Albergo et al., 2023, Liu et al., 10 Aug 2025, Hoellmer et al., 4 Feb 2025), and (Albergo et al., 2022).
1. Mathematical Definition of LSI and Theoretical Foundations
Latent Stochastic Interpolants generalize SI by defining a continuous-time stochastic process in latent space via reparameterized bridges between a fixed prior and the encoder-aggregated posterior , optimized jointly with autoencoding components. An LSI interpolant is given as
where (arbitrary prior), (learned encoder distribution), , and are scalar or vector-valued schedules parameterizing the bridge, typically satisfying
A canonical choice is the Brownian-bridge: .
The time-marginal law of interpolates between at and the aggregated posterior at exactly in finite time. This construction enables LSI to model arbitrary transformations between latent distributions, with sampling and transport realized via an SDE or its probability-flow ODE counterpart: The drift is learned to match the reference bridge determined by the interpolant.
The theoretical core is the reduction of learning a complex generative map to the problem of matching the SDE’s marginal at to the decoder’s latent aggregated posterior, with tightness of the continuous-time evidence lower bound (ELBO) when the variational drift equals the model drift (Singh et al., 2 Jun 2025, Albergo et al., 2023).
2. Variational Objective and Algorithmic Structure
The LSI variational objective is an ELBO derived explicitly in continuous time. Denoting as the path under the model SDE and as the bridge interpolation (variational posterior), the ELBO is
where is the optimal control,
with specifying the drift for the closed-form sample path defined by the bridge .
In practice, LSI is trained in a single-stage, end-to-end manner:
- Sample mini-batch , encode to stochastic , sample and noise , and time .
- Compute via the interpolant.
- Evaluate ELBO terms (reconstruction and drift loss).
- Backpropagate jointly through encoder, decoder, and .
A numerically robust variant is the parameterization, where the loss takes the form
with details in [(Singh et al., 2 Jun 2025), eq. 19] and schedule hyperparameters (e.g. ) optimized per task.
3. Numerical Solvers and Convergence Analysis
Sampling via LSI typically integrates the learned ODE/SDE in latent space. Accurate transport is guaranteed to first- or second-order in time-step for popular integrators. The TV-distance between the true and discretized path can be upper-bounded as
with first-order (Euler, ) and second-order (Heun/RK2, ) error scaling, provided is handled or early-stopping avoids endpoints where (Liu et al., 10 Aug 2025).
The interpolant noise scale determines stiffness and grid selection; adaptive schedules where substantially improve efficiency, as can employing second-order methods with only two drift evaluations per step.
4. Relation to Standard Diffusion Models and Other Generative Frameworks
LSI encompasses and generalizes standard score-based diffusion and flow-matching models:
- LSI allows arbitrary prior (Gaussian, uniform, Laplace, or learned), whereas diffusion models conventionally fix .
- LSI runs generative dynamics in a low-dimensional latent domain, sharply reducing computational cost relative to SI or diffusion ran directly in pixel space [(Singh et al., 2 Jun 2025) Table 1, Table 7].
- Both ODE-based (flow-matching) and SDE-based (score-based) LSI variants are supported by the same quadratic regression/ELBO objective, unifying flows and diffusions and connecting with Schrödinger bridge approaches (Albergo et al., 2023, Albergo et al., 2022).
- For multimarginal or multitask settings, operator- or simplex-based generalizations of the SI time index enable one-shot, multi-way LSI frameworks, supporting inpainting, style transfer, and conditional sampling without retraining (Albergo et al., 2023, Negrel et al., 6 Aug 2025).
5. Model Architectures, Parameterization, and Training Details
Empirically, in large-scale image modeling (ImageNet), the LSI pipeline comprises:
- Encoder/decoder: 2-stage convolutional stack, with up/downsampling and linear token layers.
- Latent SI (drift) network: U-Net-like stack of self-attention Transformer blocks at latent spatial resolution (e.g., ).
- Optimizer: AdamW , batch size 2048, up to 2000 epochs, learning rate warmup, cosine decay, exponential moving average with 0.9999 decay.
- Time parameterization: , possibly warped as with .
- Noise scale scanned (e.g. ), with best performance for at resolution [(Singh et al., 2 Jun 2025), Appendix I].
Latent interpolation schedules () are generally chosen as linear or Brownian-bridge-like, with stochastic encoding providing regularization and improved generative diversity.
6. Empirical Evaluation and Benchmarks
LSI demonstrates competitive or superior performance to observation-space SI and pixel-space diffusion on standard generative metrics, using order-of-magnitude fewer sampling FLOPs. On ImageNet:
- FID for LSI ("Latent") at is 3.12, outperforming SI in observation space (3.46) with reduced computational budget [(Singh et al., 2 Jun 2025), Table 1].
- Arbitrary priors (uniform, Laplace, Gaussian) deliver similar performance (e.g., FID 4.81, 4.45, 3.76) [(Singh et al., 2 Jun 2025), Table 7].
- Classifier-free guidance and inversion via SDE/ODE sampling are effective and seamlessly supported due to the flexible latent dynamics [(Singh et al., 2 Jun 2025), Figures 6 and 7].
- Physical emulation with LSI matches or improves upon FNO, DDPM, and FM baselines on PDE and climate tasks while requiring only $2$–$10$ integration steps for deterministic predictions or higher for calibrated ensemble spread (Zhou et al., 30 Sep 2025).
7. Extensions, Limitations, and Future Directions
LSI is extensible to:
- Arbitrary prior/posterior pairs, discrete and periodic-valued spaces (e.g., with atomistic flows in crystalline materials (Hoellmer et al., 4 Feb 2025)), operator-based multitask learning (Negrel et al., 6 Aug 2025), and multimarginal translation (Albergo et al., 2023).
- SDE or ODE solvers, with practical guidance to handle endpoint singularities or hyperparameter sensitivity.
- Multimodal, hierarchical, or domain-conditional latent representations (Singh et al., 2 Jun 2025).
Limitations noted include:
- Variational posterior is often assumed to follow linear bridges and fixed , potentially restricting expressiveness relative to general diffusion bridges.
- Numerical issues near due to vanishing (endpoint singularity).
- Quantitative performance is coupled to expressive power of the autoencoder/decoder and capacity of the drift network.
- Theoretical guarantees assume correctly specified conditional expectations and sufficiently expressive function approximators.
Open research directions include learning more flexible (-dependent, nonlinear) variational bridges, extension to other modalities (audio, video), tighter ELBO variants, explicit connections with Schrödinger bridge or Wasserstein gradient flows in learned latent geometries, and adaptive numerical integration strategies to further optimize sampling tradeoffs (Singh et al., 2 Jun 2025).
References:
- "Latent Stochastic Interpolants" (Singh et al., 2 Jun 2025)
- "Stochastic Interpolants: A Unifying Framework for Flows and Diffusions" (Albergo et al., 2023)
- "Finite-Time Convergence Analysis of ODE-based Generative Models for Stochastic Interpolants" (Liu et al., 10 Aug 2025)
- "Open Materials Generation with Stochastic Interpolants" (Hoellmer et al., 4 Feb 2025)
- "Building Normalizing Flows with Stochastic Interpolants" (Albergo et al., 2022)
- "Multimarginal generative modeling with stochastic interpolants" (Albergo et al., 2023)
- "Multitask Learning with Stochastic Interpolants" (Negrel et al., 6 Aug 2025)
- "Reframing Generative Models for Physical Systems using Stochastic Interpolants" (Zhou et al., 30 Sep 2025)