SCLD: Controlled Langevin Diffusion
- SCLD is a framework that integrates sequential Monte Carlo with controlled Langevin dynamics to sample from unnormalized target densities.
- It employs a continuous-time path-space formulation with learnable drift functions, systematic resampling, and log-variance loss to control bias and variance.
- Empirical benchmarks demonstrate that SCLD achieves efficient, high-quality sampling in complex, high-dimensional settings while significantly reducing computational cost.
Sequential Controlled Langevin Diffusion (SCLD) is a principled framework for sampling from unnormalized target densities by synthesizing the asymptotic robustness of Sequential Monte Carlo (SMC) with the flexibility and adaptivity of diffusion-based sampling methods. SCLD introduces a continuous-time path-space perspective, leveraging learnable Langevin drifts and systematic resampling to achieve efficient, high-quality sampling in complex and high-dimensional distributions, with significant reduction in computational cost compared to traditional diffusion samplers (Chen et al., 2024).
1. Problem Setting and Interpolating Distributions
The fundamental objective in SCLD is to generate samples from an unnormalized density
where is tractable but %%%%2%%%% is unknown. SCLD introduces a continuous-time annealing process via the interpolant
such that (typically Gaussian) and . The schedule function can be linear or learned, providing a smooth progression from easy-to-sample priors to the complex target (Chen et al., 2024).
2. Path-Space Formulation and SDEs
SCLD models the annealing process on path space using controlled Langevin stochastic differential equations (SDEs):
- Forward SDE:
where is a learnable time-dependent drift.
- Backward SDE (via Nelson’s identity):
The unique optimal control equalizes the forward and reverse path measures, representing the solution to the continuous-time Schrödinger bridge (Chen et al., 2024).
3. Variational Objective and Log-Variance Loss
The learning problem for the drift is formulated via a variational objective on path space. Let and $\cev P^b$ denote the forward and reverse SDE path laws over . The Radon–Nikodym derivative (RND) between reverse and forward path measures is
$w(X_{[0, T]}) = \frac{\mathrm{d} \cev P^b}{\mathrm{d} \vec P^b}(X_{[0,T]}) = \exp\Bigl(\log \pi(X_T, T) - \log \pi(X_0, 0) - \textstyle{\tfrac{1}{2}} \int_0^T \|b_s - 2\nabla \log \pi\|^2 - \|b_s\|^2\,\mathrm{d}s + \cdots\Bigr).$
One minimizes variational losses such as $\mathrm{KL}(\vec P^b \| \cev P^b) = - \mathbb{E}_{\vec P^b}[\log w]$ or the log-variance loss $\mathrm{LV}(\vec P^b\|\cev P^b) = \mathrm{Var}_{\vec P^b}[\log w]$. Empirically the log-variance loss exhibits superior scaling (polynomial, not exponential variance) in high dimension (Chen et al., 2024).
4. Algorithmic Construction: SMC Integration and Discretization
SCLD divides into subintervals . For each, the RND over the interval is defined, allowing the total path weight to factorize:
Particles are propagated forward following discretized Euler–Maruyama steps:
with step size for subinterval length . Weights are updated via the discretized RND, and resampling is triggered whenever the effective sample size (ESS) falls below a fixed threshold (e.g., $0.3K$). Optionally, a few MCMC steps targeting are applied after resampling to enhance mixing (Chen et al., 2024).
Pseudocode for SCLD is as follows:
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 |
Input: prior p, target ρ, schedule β, times 0=t₀<…<t_N=T,
substeps L, particles K, drift-net b_θ
Initialize X₀^{(k)}∼p, w₀^{(k)}=1 for k=1…K
for n=1…N do
for i=1…L do
t←t_{n-1}+i*(t_n−t_{n-1})/L
X_i^{(k)}←X_{i-1}^{(k)}+b_θ(X_{i-1}^{(k)},t−h)*h+√(2h)*ξ
end
compute π(·,t_n) and w_{[t_{n-1},t_n]}^{(k)} via discrete RND
w_n^{(k)}←w_{n-1}^{(k)}·w_{[t_{n-1},t_n]}^{(k)}
if ESS({w_n^{(k)})<threshold
resample {X_{nL}^{(k)},w_n^{(k)}→{X̃^{(k)},1/K}
optional: one MCMC step targeting π(·,t_n)
end
end
Return {X_T^{(k)}≈p_target} |
5. Theoretical Properties and Bias–Variance Tradeoff
If the drift exactly solves the path-space Schrödinger bridge, then all path weights become unity and no resampling is required, yielding perfect finite-time transport. In practice, unbiasedness for expectations under is retained. Theoretical analysis establishes that the variance of the KL-loss-based estimator grows exponentially with dimension when relying on interval reweighting, but the log-variance loss reduces this scaling to polynomial, which is critical for high-dimensional efficiency (Proposition 5.7 in (Chen et al., 2024)). Discretization bias can be controlled by adjusting the Euler–Maruyama step size, trading computation for accuracy.
6. Empirical Performance, Robustness, and Benchmarks
SCLD has been evaluated across 11 benchmarks, covering Bayesian logistic and random-effect models (up to 1,600 dimensions), synthetic multimodal Gaussian mixtures (up to 40 modes in 50 dimensions), and robotic-arm motion planning with sharp, separated modes. Evaluation metrics include ELBO (for estimation of ) when is unknown, and Sinkhorn–Wasserstein distance to ground-truth samples for others.
Empirically, SCLD attains or surpasses the performance of classical SMC, CRAFT, DDS, PIS, and CMCD variants, often requiring only 10% of the gradient computation budget of pure diffusion-based samplers (e.g., SCLD achieves convergence with approximately 3,000 steps vs. 40,000). Convergence rate improvements of 5–10x in ELBO wall-clock time are observed relative to ULA-based SMC (Chen et al., 2024). The robustness of SCLD is attributed to synergistic use of log-variance loss, replay buffer, systematic resampling, and occasional MCMC corrections; ablation studies confirm the necessity of each component.
7. Relationship to Prior Sequential Monte Carlo and Langevin Approaches
The SCLD method unifies principles from both SMC—such as particle propagation, adaptive resampling, and importance weighting—and controlled Langevin dynamics, in the sense of learned, time-dependent drift functions governing the proposals. Earlier approaches (e.g., the controlled-Langevin SMC model of (Septier et al., 2015)) adopted discrete-time, stepwise controlled Langevin mutation for Bayesian filtering, combining mutation, weighting, and Metropolized-Langevin moves. SCLD distinguishes itself by explicitly casting the problem in continuous time, learning the drift to optimize path-space transport, and integrating a low-variance loss on trajectories. Both frameworks employ effective sample size to guide resampling and deploy flexible drift adjustments to counteract sample impoverishment in high dimensions. However, SCLD's use of a learnable drift network and variational path-space loss provides additional adaptivity and computational efficiency (Chen et al., 2024, Septier et al., 2015).