Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
173 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Skew-Reflected Non-Reversible Langevin Dynamics

Updated 30 June 2025
  • SRNLD is a stochastic process that enhances sampling from constrained domains by combining non-reversible antisymmetric drift with skewed boundary reflections.
  • It rotates the gradient flow using a skew-symmetric matrix to accelerate convergence and reduce autocorrelation while preserving the invariant measure.
  • SRNLD offers rigorous convergence guarantees and has shown improved performance in applications such as Bayesian regression and high-dimensional classification.

Skew-Reflected Non-Reversible Langevin Dynamics (SRNLD) is a class of stochastic processes designed for efficient sampling from target distributions supported on constrained domains. SRNLD generalizes classical reversible reflected Langevin dynamics by introducing non-reversibility through antisymmetric drift and constructing a "skew" boundary reflection mechanism. This combination yields accelerated convergence toward equilibrium and sharper concentration of the process around its invariant measure, especially in high-dimensional or complex constrained settings.

1. Mathematical Structure and Motivation

SRNLD aims to sample from a probability density μ(x)exp(f(x))\mu(x) \propto \exp(-f(x)) defined on a constrained domain KRdK \subset \mathbb{R}^d, where f:RdRf: \mathbb{R}^d \to \mathbb{R} is smooth. The continuous-time dynamics are given by the skew-reflected stochastic differential equation: dXt=(I+J(Xt))f(Xt)dt+2dWt+nJ(Xt)L(dt)dX_t = - (I + J(X_t)) \nabla f(X_t) dt + \sqrt{2} dW_t + \mathbf{n}^J(X_t) L(dt) where

  • J(x)J(x) is a skew-symmetric matrix field (J(x)=J(x)J(x) = -J(x)^\top),
  • WtW_t is a standard Brownian motion,
  • nJ(x)=(I+J(x))n(x)n(x)22+J(x)n(x)22\mathbf{n}^J(x) = \frac{(I + J(x))\mathbf{n}(x)}{\sqrt{\|\mathbf{n}(x)\|_2^2 + \|J(x)\mathbf{n}(x)\|_2^2}} adjusts the normal vector at the boundary K\partial K to ensure the process remains in KK,
  • L(dt)L(dt) is a measure increasing only when XtKX_t \in \partial K, i.e., local time on the boundary,
  • n(x)\mathbf{n}(x) is the inward unit normal at xKx \in \partial K.

The innovation in SRNLD is the coordinated design of the antisymmetric drift J(x)J(x) and the "skew" boundary reflection to maintain the desired invariant measure while breaking detailed balance and improving ergodic properties.

2. Non-Reversibility and Skew Reflection Mechanism

The non-reversible drift is introduced via the skew-symmetric J(x)J(x), resulting in a drift vector

(I+J(x))f(x)-(I + J(x))\nabla f(x)

which rotates the gradient flow. This non-gradient perturbation induces probability currents within KK, breaking reversibility (detailed balance), a property which—when properly constructed—does not alter the invariant measure but accelerates mixing and reduces sampling autocorrelation.

The boundary reflection is "skew" in the sense that, when the process hits K\partial K, it is reflected in the direction (I+J(x))n(x)(I + J(x))\mathbf{n}(x), not simply n(x)\mathbf{n}(x). This ensures compatibility with the non-reversible interior drift and maintains the target density as invariant for the process. The mathematical justification and well-posedness are established via an associated Skorokhod problem with oblique (skewed) reflection; under convexity and regularity conditions on KK and smoothness of J(x)J(x), existence and uniqueness hold.

3. Boundary Conditions and Design of the Skew-Symmetric Matrix

The construction of an admissible matrix field J(x)J(x) is crucial. Analysis shows that to retain the invariant measure and guarantee well-posedness, it is necessary that

J(x)n(x)=0for all xK.J(x)\mathbf{n}(x) = 0 \qquad \text{for all } x \in \partial K.

This requirement ensures the skew reflection does not "push" the process outside the domain and allows the boundary condition for the generator to reduce to a Neumann-type form, facilitating both theoretical analysis and implementation.

Within the interior KK^\circ, further divergence-free constraints (J(x)=0\nabla \cdot J(x) = 0) may be imposed for the process to admit the correct stationary law.

Construction examples include:

  • For the unit ball, J(x)J(x) may be constructed via a block-diagonal or cross-product form ensuring J(x)x=0J(x)x = 0 for xKx \in \partial K.
  • For general convex domains, J(x)J(x) is built to annihilate the local normal direction at the boundary.

The practical construction of efficient J(x)J(x) is informed by domain geometry and the need to maintain ergodicity and reflection consistency.

4. Convergence Guarantees and Large Deviations Theory

SRNLD achieves strictly faster convergence to the target law compared to its reversible counterpart. Let ρJ\rho_J denote the spectral gap associated with the generator of SRNLD, and ρ0\rho_0 that of the reversible RLD. Then,

ρJρ0>0,\rho_J \geq \rho_0 > 0,

implying exponential convergence in total variation and $1$-Wasserstein distance: TV(Law(Xt),μ)KeρJt,W1(Law(Xt),μ)2RKeρJt,\mathrm{TV}(\mathrm{Law}(X_t), \mu) \leq \mathcal{K} e^{-\rho_J t}, \qquad \mathcal{W}_1(\mathrm{Law}(X_t), \mu) \leq 2R \cdot \mathcal{K} e^{-\rho_J t}, where RR bounds KK.

Further, a large deviation principle (LDP) for the empirical measures of SRNLD reveals enhanced concentration relative to the reversible process. The LDP rate function decomposes as: I(ν)=IS(ν)+IA(ν)I(\nu) = I_S(\nu) + I_A(\nu) with:

  • IS(ν)I_S(\nu) corresponding to the symmetric (reversible) part,
  • IA(ν)=14J(x)f(x)v(x)HS1(ν)2I_A(\nu) = \frac{1}{4} \| -J(x)\nabla f(x)\cdot\nabla v(x) \|_{\mathcal{H}^{-1}_S(\nu)}^2 a non-negative contribution due to the antisymmetric drift, where v=logdνdμv = \log \frac{d\nu}{d\mu}.

Since IA(ν)0I_A(\nu)\geq 0, large deviations are rarer in SRNLD than in RLD—all else equal—quantifying a fundamental acceleration in sampling.

5. Discretization: Skew-Reflected Non-Reversible Langevin Monte Carlo

To enable implementation, SRNLD can be discretized using Euler–Maruyama-type schemes. The Skew-Reflected Non-Reversible Langevin Monte Carlo (SRNLMC) update is: xk+1=PKJ[xkη(I+J(xk))f(xk)+2ηξk+1]x_{k+1} = \mathcal{P}^J_K \Big[x_k - \eta (I + J(x_k))\nabla f(x_k) + \sqrt{2\eta} \xi_{k+1}\Big] where:

  • ξk+1N(0,I)\xi_{k+1} \sim \mathcal{N}(0, I),
  • PKJ\mathcal{P}^J_K is the skew projection along the appropriately modified normal.

Non-asymptotic discretization error bounds are established. With step size η\eta and KK steps, the overall sampling error in $1$-Wasserstein distance is controlled as

W1(Law(xK),μ)2RKeρJKη+(discretization error),\mathcal{W}_1(\mathrm{Law}(x_K), \mu) \leq 2R\mathcal{K} e^{-\rho_J K\eta} + \text{(discretization error)},

where the iteration complexity improves as ρJ\rho_J increases with better design of JJ.

6. Empirical Performance and Applications

Experimental studies confirm that SRNLMC with properly constructed J(x)J(x) outperforms projected/reversible Langevin algorithms on a variety of sampling tasks including:

  • Sampling truncated multivariate normals in balls and cubes,
  • Bayesian linear and logistic regression with norm constraints,
  • Real-world data classification tasks with high-dimensional parameter constraints.

Performance metrics—such as convergence in Wasserstein distance, mean squared error in posterior estimation, and classification accuracy—consistently validate the theoretical predictions. The advantage is most pronounced when SRNLMC utilizes a state-dependent J(x)J(x) structured to vanish along the normal at the boundary, which both satisfies the theoretical requirements and yields robust and stable practical behavior.

A summary table highlighting the principal distinctions:

Algorithm Boundary Reflection Reversibility Convergence Rate Boundary Condition for J(x)J(x)
PLMC Standard (Neumann) Reversible ρ0\rho_0 none
SRNLMC Skew (oblique) Non-reversible ρJρ0\rho_J \geq \rho_0 J(x)n(x)=0J(x)\mathbf{n}(x) = 0 on K\partial K

7. Theoretical Significance and Future Directions

SRNLD connects and generalizes key principles in modern sampling theory:

  • It leverages the acceleration properties of non-reversible dynamics known from unconstrained settings,
  • Transfers these gains to constrained domains via careful modification of both drift and boundary interaction,
  • Is grounded in large deviations and spectral theory, providing rigorous quantification of sampling acceleration,
  • Guides practical algorithm design through precise conditions on the antisymmetric matrix.

Open questions include optimal construction of J(x)J(x) for more general or non-convex sets, extensions to additional constraint types, and integration into large-scale machine learning workflows.


References:

  • (1103.2845) describes partially elastic reflected Langevin processes and their connection to skew reflections and non-reversibility.
  • (2501.11743, 2506.07816) provide mathematical formulation, convergence analysis, and large deviations results for SRNLD on convex constrained domains, including practical construction and performance benchmarks.
  • Additional context on theory and methods is drawn from the cited works in each original abstract.