Papers
Topics
Authors
Recent
2000 character limit reached

Hessian Riemannian Langevin Monte Carlo (HRLMC)

Updated 19 November 2025
  • HRLMC is a geometric sampling algorithm that extends Langevin Monte Carlo using Hessian manifolds and mirror descent for high-dimensional, non-Euclidean problems.
  • The algorithm provides non-asymptotic convergence bounds in Wasserstein and KL divergences by accommodating local metric properties and constrained geometries.
  • By unifying Riemannian SDEs, interior-point methods, and optimal transport, HRLMC offers robust performance and explicit metric-dependent contraction rates.

The Hessian Riemannian Langevin Monte Carlo (HRLMC) algorithm is a geometric extension of Langevin Monte Carlo for efficient sampling from high-dimensional probability densities, particularly those lacking global regularity in the standard Euclidean setting. HRLMC leverages a Hessian manifold structure and mirror-descent techniques to model non-flat or constrained geometries, providing non-asymptotic convergence bounds in Wasserstein and Kullback–Leibler distances. This framework unifies Riemannian stochastic differential equations (SDEs), interior-point methods, and optimal transport theory, resolving difficulties in the analysis and computation of Riemannian discretizations by introducing relative smoothness, self-concordance, and explicit metric-dependent contraction rates (Zhang et al., 2020, Gatmiry et al., 2022).

1. Hessian Manifold Structure and Riemannian Metrics

Let XRpX \subset \mathbb{R}^p be open and contractible, and let φ:XR\varphi : X \to \mathbb{R} be a C2C^2 Legendre-type convex function (often called the "mirror potential" or barrier function). The Hessian manifold is endowed with the metric

gx(u,v)=u[D2φ(x)]vg_x(u, v) = u^\top [D^2 \varphi(x)]\, v

where D2φ(x)D^2 \varphi(x) is positive definite for all xXx \in X. This construction allows dual coordinates y=φ(x)y = \nabla \varphi(x), in which the metric becomes Euclidean and Riemannian quantities push forward via the mirror map. Important classes of φ\varphi include the Euclidean case φ(x)=12x2\varphi(x) = \frac{1}{2} \|x\|^2 and the log-barrier φ(x)=ilog(aixbi)\varphi(x) = -\sum_i \log(a_i^\top x - b_i) for polytope constraints, which yield g(x)=2φ(x)g(x) = \nabla^2 \varphi(x) (Zhang et al., 2020, Gatmiry et al., 2022).

2. Riemannian Langevin Diffusion: SDE Formulation

Let π\pi be a target density dπ(x)=ef(x)dxd\pi(x) = e^{-f(x)} dx, fC3(X)f \in C^3(X). The Riemannian Langevin diffusion (RLD) on (X,g)(X, g) is the Itô SDE:

dXt=[θ(Xt)D2φ(Xt)1f(Xt)]dt+2  D2φ(Xt)1/2dBtdX_t = \Big[\theta(X_t) - D^2\varphi(X_t)^{-1} \nabla f(X_t)\Big]dt + \sqrt{2}\; D^2\varphi(X_t)^{-1/2} dB_t

where

θ(x)=D2φ(x)1Tr[D3φ(x)D2φ(x)1]\theta(x) = -D^2\varphi(x)^{-1} \, \mathrm{Tr}[ D^3 \varphi(x) D^2\varphi(x)^{-1} ]

The invariant measure of this diffusion is π\pi by Fokker–Planck analysis. In coordinates y=φ(x)y = \nabla\varphi(x) the SDE simplifies, and the diffusion is natural Brownian motion on the manifold, with drift g1(logν)g^{-1}\nabla (\log \nu) and diffusion g1/2g^{-1/2} (Zhang et al., 2020, Gatmiry et al., 2022).

3. Discrete HRLMC Scheme and Mirror Descent Analogy

The discretization leading to HRLMC is performed in the dual space and mapped back via the Legendre conjugate. With step-size hk+1h_{k+1} and i.i.d. ξk+1N(0,Ip)\xi_{k+1} \sim \mathcal{N}(0, I_p): Initialize x0X yk=φ(xk) yk+1=ykhk+1f(xk)+2hk+1D2φ(xk)ξk+1 xk+1=φ(yk+1)\begin{align*} \text{Initialize } x_0 \in X \ y_k = \nabla \varphi(x_k) \ y_{k+1} = y_k - h_{k+1} \nabla f(x_k) + \sqrt{2 h_{k+1} D^2\varphi(x_k)}\, \xi_{k+1} \ x_{k+1} = \nabla \varphi^*(y_{k+1}) \end{align*} HRLMC is the sampling analogue of mirror descent: setting ξ=0\xi = 0 yields the mirror descent step for ff with mirror map φ\varphi. The stochastic perturbation ensures the correct stationary distribution π\pi. In the Euclidean case, HRLMC recovers Unadjusted Langevin Algorithm (ULA) (Zhang et al., 2020). For constrained domains (e.g., polytopes), ULA in local charts uses the metric and steps: xk+1=xkη  g(xk)1f(xk)+2η  g(xk)1/2ξkx_{k+1} = x_k - \eta\; g(x_k)^{-1} \nabla f(x_k) + \sqrt{2\eta}\; g(x_k)^{-1/2} \xi_k (Gatmiry et al., 2022).

4. Assumptions, Metric Properties, and Self-Concordance

Non-asymptotic convergence proofs require the following conditions:

  • A1 (Self-concordance-like control): κ0\exists \kappa \geq 0 such that D2φ(x)1/2D2φ(x)1/2Fκφ(x)φ(x)2\|D^2\varphi(x)^{1/2} - D^2\varphi(x')^{1/2}\|_F \leq \kappa \|\nabla\varphi(x) - \nabla\varphi(x')\|_2.
  • A2 (Moment bound): R=EXπ[D2φ(X)2]<R = \mathbb{E}_{X \sim \pi}[\|D^2\varphi(X)\|_2] < \infty.
  • A3 (Relative strong convexity of ff): m0\exists m \geq 0 with mφ(x)φ(x)22f(x)f(x),φ(x)φ(x)m\|\nabla\varphi(x) - \nabla\varphi(x')\|_2^2 \leq \langle\nabla f(x)-\nabla f(x'), \nabla\varphi(x)-\nabla\varphi(x')\rangle.
  • A4 (Relative gradient Lipschitzness): M>0\exists M > 0 such that f(x)f(x)2Mφ(x)φ(x)2\|\nabla f(x) - \nabla f(x')\|_2 \leq M \|\nabla\varphi(x) - \nabla\varphi(x')\|_2.
  • A5 (Commutator bound): [D2φ(x)1,D2f(x)]2δ\|\left[D^{2}\varphi(x)^{-1}, D^2f(x)\right]\|_2 \leq \delta.

Self-concordance is extended to second-order conditions: bounds on derivatives of the metric and its inverse, controlling the Jacobian growth and ensuring discretization errors remain manageable (Gatmiry et al., 2022).

5. Non-Asymptotic Convergence Bounds and Wasserstein Geometry

HRLMC delivers explicit, non-asymptotic contraction rates in the Wasserstein ground metric: W2,φ2(μ,ν)=infXμ,XνE[φ(X)φ(X)22]W_{2,\varphi}^2(\mu, \nu) = \inf_{X \sim \mu, X' \sim \nu} \mathbb{E} \left[\|\nabla\varphi(X) - \nabla\varphi(X')\|_2^2\right] With step-size hk+1h_{k+1} and contraction factor ρk+1\rho_{k+1}, for all kk,

W2,φ(μk+1,π)ρk+1W2,φ(μk,π)+hk+1p1/2β1(R,κ)+hk+13/2p1/2β2(R,M,κ)W_{2,\varphi}(\mu_{k+1}, \pi) \leq \rho_{k+1}\, W_{2,\varphi}(\mu_k, \pi) + h_{k+1}\, p^{1/2}\, \beta_1(R,\kappa) + h_{k+1}^{3/2} p^{1/2} \beta_2(R,M,\kappa)

where κ~=κ2+δ(4M+δ)/2(m+M)\tilde{\kappa} = \sqrt{\kappa^2 + \delta (4M+\delta)/2(m+M)} and explicit constants β1,β2\beta_1, \beta_2 are supplied as functions of dimension, moment bounds, and regularity parameters. For constant step-size, iterates contract to an explicit Wasserstein ball around π\pi and remain inside, with radius proportional to κ\kappa and local geometry (Zhang et al., 2020).

For second-order self-concordant metrics, the convergence in Kullback–Leibler divergence is: Hν(ρk)ecαϵkHν(ρ0)+O(n5/2L2+nL22+nL3)ϵαH_\nu(\rho_k) \leq e^{-c\,\alpha\,\epsilon\, k} H_\nu(\rho_0) + O(n^{5/2} L_2 + \sqrt{n} L_2^2 + n L_3) \frac{\epsilon}{\alpha} where HνH_\nu denotes KL divergence relative to ν\nu and L2L_2, L3L_3 are Lipschitz constants of FF and its Hessian (Gatmiry et al., 2022).

6. Specializations and Applications

Euclidean LMC: Setting φ(x)=12x2\varphi(x) = \frac{1}{2}\|x\|^2 yields g=Ig = I, κ=δ=0\kappa = \delta = 0, recovers classical LMC with optimal Wasserstein bounds.

Constrained Sampling: For polytopes M={x:Axb}\mathcal{M} = \{x: Ax \geq b\}, using the log-barrier φ(x)=ilog(aixbi)\varphi(x) = -\sum_i \log(a_i^\top x - b_i) yields g(x)=2φ(x)g(x) = \nabla^2 \varphi(x) with self-concordant constants γ1=2,γ2=4,γ3=6\gamma_1 = 2, \gamma_2 = 4, \gamma_3 = 6. HRLMC enables polynomial-time sampling of isoperimetric densities given gradient oracles and avoids global smoothness requirements.

Other Metrics: Adapting φ\varphi to the geometry of ff allows sampling from densities in non-Euclidean geometries, e.g., Dirichlet or gamma posteriors on the simplex (Zhang et al., 2020).

7. Key Analytical Tools and Proofs

HRLMC analysis relies on:

  • Extended Baillon–Haddad Inequality: For relative smoothness, establishes cocoercivity bounds between gradients of ff and the mirror map.
  • Itô’s Isometry and Minkowski Integral Inequality: Manage stochastic error terms and bias in diffusion discretization.
  • Chung’s Lemma: Demonstrates contraction to a Wasserstein ball under vanishing step-sizes.
  • Bismut–Elworthy–Li Formula: Controls higher-order terms in stochastic flows.
  • Self-Concordance Theory: Manages metric derivatives and stability (Nesterov–Nemirovskii).

These tools together establish dimension-dependent rates, validate non-asymptotic guarantees, and ensure robustness to variance in local geometry (Zhang et al., 2020, Gatmiry et al., 2022).


HRLMC thus constitutes an advanced geometric Monte Carlo method tailored for high-dimensional, non-Euclidean, and constrained sampling problems, bridging stochastic analysis, convex optimization, and differential geometry. The rigorous contraction rates and metric-dependent bounds enable its application to non-smooth log-concave target distributions and provide theoretical foundations for sampling in complex domains.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (2)
Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Hessian Riemannian Langevin Monte Carlo (HRLMC) Algorithm.