Papers
Topics
Authors
Recent
Search
2000 character limit reached

SCLD: Controlled Langevin Diffusion

Updated 28 January 2026
  • SCLD is a framework that integrates sequential Monte Carlo with controlled Langevin dynamics to sample from unnormalized target densities.
  • It employs a continuous-time path-space formulation with learnable drift functions, systematic resampling, and log-variance loss to control bias and variance.
  • Empirical benchmarks demonstrate that SCLD achieves efficient, high-quality sampling in complex, high-dimensional settings while significantly reducing computational cost.

Sequential Controlled Langevin Diffusion (SCLD) is a principled framework for sampling from unnormalized target densities by synthesizing the asymptotic robustness of Sequential Monte Carlo (SMC) with the flexibility and adaptivity of diffusion-based sampling methods. SCLD introduces a continuous-time path-space perspective, leveraging learnable Langevin drifts and systematic resampling to achieve efficient, high-quality sampling in complex and high-dimensional distributions, with significant reduction in computational cost compared to traditional diffusion samplers (Chen et al., 2024).

1. Problem Setting and Interpolating Distributions

The fundamental objective in SCLD is to generate samples from an unnormalized density

ptarget(x)=ρtarget(x)Z,Z=Rdρtarget(x)dx,p_{\rm target}(x) = \frac{\rho_{\rm target}(x)}{Z}, \qquad Z = \int_{\mathbb{R}^d} \rho_{\rm target}(x)\,\mathrm{d}x,

where ρtarget(x)\rho_{\rm target}(x) is tractable but %%%%2%%%% is unknown. SCLD introduces a continuous-time annealing process via the interpolant

π(x,t)pprior(x)1β(t)ρtarget(x)β(t),t[0,T],β(0)=0,  β(T)=1,\pi(x,t) \propto p_{\rm prior}(x)^{1-\beta(t)}\,\rho_{\rm target}(x)^{\beta(t)}, \qquad t \in [0, T],\quad \beta(0)=0,\;\beta(T)=1,

such that π(,0)=pprior\pi(\cdot, 0) = p_{\rm prior} (typically Gaussian) and π(,T)=ptarget\pi(\cdot, T) = p_{\rm target}. The schedule function β(t)\beta(t) can be linear or learned, providing a smooth progression from easy-to-sample priors to the complex target (Chen et al., 2024).

2. Path-Space Formulation and SDEs

SCLD models the annealing process on path space using controlled Langevin stochastic differential equations (SDEs):

dXt=bt(Xt)dt+2dWt,X0pprior\mathrm{d}X_t = b_t(X_t)\,\mathrm{d}t + \sqrt{2}\,\mathrm{d}W_t, \qquad X_0 \sim p_{\rm prior}

where bt:RdRdb_t:\mathbb{R}^d \to \mathbb{R}^d is a learnable time-dependent drift.

  • Backward SDE (via Nelson’s identity):

dYt=(bt(Yt)2logπ(Yt,t))dt+2dW~t,YTptarget\mathrm{d}Y_t = \bigl(b_t(Y_t) - 2 \nabla \log \pi(Y_t, t)\bigr)\,\mathrm{d}t + \sqrt{2}\,\mathrm{d}\widetilde W_t, \qquad Y_T \sim p_{\rm target}

The unique optimal control bb^* equalizes the forward and reverse path measures, representing the solution to the continuous-time Schrödinger bridge (Chen et al., 2024).

3. Variational Objective and Log-Variance Loss

The learning problem for the drift bb is formulated via a variational objective on path space. Let Pb\vec P^b and $\cev P^b$ denote the forward and reverse SDE path laws over t[0,T]t \in [0, T]. The Radon–Nikodym derivative (RND) between reverse and forward path measures is

$w(X_{[0, T]}) = \frac{\mathrm{d} \cev P^b}{\mathrm{d} \vec P^b}(X_{[0,T]}) = \exp\Bigl(\log \pi(X_T, T) - \log \pi(X_0, 0) - \textstyle{\tfrac{1}{2}} \int_0^T \|b_s - 2\nabla \log \pi\|^2 - \|b_s\|^2\,\mathrm{d}s + \cdots\Bigr).$

One minimizes variational losses such as $\mathrm{KL}(\vec P^b \| \cev P^b) = - \mathbb{E}_{\vec P^b}[\log w]$ or the log-variance loss $\mathrm{LV}(\vec P^b\|\cev P^b) = \mathrm{Var}_{\vec P^b}[\log w]$. Empirically the log-variance loss exhibits superior scaling (polynomial, not exponential variance) in high dimension (Chen et al., 2024).

4. Algorithmic Construction: SMC Integration and Discretization

SCLD divides [0,T][0, T] into NN subintervals [tn1,tn][t_{n-1}, t_n]. For each, the RND over the interval is defined, allowing the total path weight to factorize:

w[0,T]=n=1Nw[tn1,tn].w_{[0, T]} = \prod_{n=1}^N w_{[t_{n-1}, t_n]}.

Particles are propagated forward following discretized Euler–Maruyama steps:

Xi=Xi1+bti1(Xi1)h+2h  ξi,ξiN(0,I)X_i = X_{i-1} + b_{t_{i-1}}(X_{i-1})\,h + \sqrt{2h}\;\xi_i, \qquad \xi_i \sim \mathcal{N}(0, I)

with step size h=τ/Lh = \tau/L for subinterval length τ\tau. Weights are updated via the discretized RND, and resampling is triggered whenever the effective sample size (ESS) ESS=(kwk)2/kwk2\mathrm{ESS} = (\sum_k w_k)^2 / \sum_k w_k^2 falls below a fixed threshold (e.g., $0.3K$). Optionally, a few MCMC steps targeting π(,tn)\pi(\cdot, t_n) are applied after resampling to enhance mixing (Chen et al., 2024).

Pseudocode for SCLD is as follows:

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
Input: prior p, target ρ, schedule β, times 0=t₀<…<t_N=T,
       substeps L, particles K, drift-net b_θ

Initialize X₀^{(k)}∼p, w₀^{(k)}=1 for k=1…K
for n=1…N do
  for i=1…L do
    t←t_{n-1}+i*(t_n−t_{n-1})/L
    X_i^{(k)}←X_{i-1}^{(k)}+b_θ(X_{i-1}^{(k)},t−h)*h+√(2h)*ξ
  end
  compute π(·,t_n) and w_{[t_{n-1},t_n]}^{(k)} via discrete RND
  w_n^{(k)}←w_{n-1}^{(k)}·w_{[t_{n-1},t_n]}^{(k)}
  if ESS({w_n^{(k)})<threshold
    resample {X_{nL}^{(k)},w_n^{(k)}→{X̃^{(k)},1/K}
    optional: one MCMC step targeting π(·,t_n)
  end
end

Return {X_T^{(k)}≈p_target}
(Chen et al., 2024)

5. Theoretical Properties and Bias–Variance Tradeoff

If the drift bb exactly solves the path-space Schrödinger bridge, then all path weights ww become unity and no resampling is required, yielding perfect finite-time transport. In practice, unbiasedness for expectations under ptargetp_{\rm target} is retained. Theoretical analysis establishes that the variance of the KL-loss-based estimator grows exponentially with dimension when relying on interval reweighting, but the log-variance loss reduces this scaling to polynomial, which is critical for high-dimensional efficiency (Proposition 5.7 in (Chen et al., 2024)). Discretization bias can be controlled by adjusting the Euler–Maruyama step size, trading computation for accuracy.

6. Empirical Performance, Robustness, and Benchmarks

SCLD has been evaluated across 11 benchmarks, covering Bayesian logistic and random-effect models (up to 1,600 dimensions), synthetic multimodal Gaussian mixtures (up to 40 modes in 50 dimensions), and robotic-arm motion planning with sharp, separated modes. Evaluation metrics include ELBO (for estimation of logZ\log Z) when ZZ is unknown, and Sinkhorn–Wasserstein distance to ground-truth samples for others.

Empirically, SCLD attains or surpasses the performance of classical SMC, CRAFT, DDS, PIS, and CMCD variants, often requiring only 10% of the gradient computation budget of pure diffusion-based samplers (e.g., SCLD achieves convergence with approximately 3,000 steps vs. 40,000). Convergence rate improvements of 5–10x in ELBO wall-clock time are observed relative to ULA-based SMC (Chen et al., 2024). The robustness of SCLD is attributed to synergistic use of log-variance loss, replay buffer, systematic resampling, and occasional MCMC corrections; ablation studies confirm the necessity of each component.

7. Relationship to Prior Sequential Monte Carlo and Langevin Approaches

The SCLD method unifies principles from both SMC—such as particle propagation, adaptive resampling, and importance weighting—and controlled Langevin dynamics, in the sense of learned, time-dependent drift functions governing the proposals. Earlier approaches (e.g., the controlled-Langevin SMC model of (Septier et al., 2015)) adopted discrete-time, stepwise controlled Langevin mutation for Bayesian filtering, combining mutation, weighting, and Metropolized-Langevin moves. SCLD distinguishes itself by explicitly casting the problem in continuous time, learning the drift to optimize path-space transport, and integrating a low-variance loss on trajectories. Both frameworks employ effective sample size to guide resampling and deploy flexible drift adjustments to counteract sample impoverishment in high dimensions. However, SCLD's use of a learnable drift network and variational path-space loss provides additional adaptivity and computational efficiency (Chen et al., 2024, Septier et al., 2015).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (2)

Topic to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Sequential Controlled Langevin Diffusion (SCLD).