Hessian Riemannian Langevin Monte Carlo (HRLMC)
- HRLMC is a geometric sampling algorithm that extends Langevin Monte Carlo using Hessian manifolds and mirror descent for high-dimensional, non-Euclidean problems.
- The algorithm provides non-asymptotic convergence bounds in Wasserstein and KL divergences by accommodating local metric properties and constrained geometries.
- By unifying Riemannian SDEs, interior-point methods, and optimal transport, HRLMC offers robust performance and explicit metric-dependent contraction rates.
The Hessian Riemannian Langevin Monte Carlo (HRLMC) algorithm is a geometric extension of Langevin Monte Carlo for efficient sampling from high-dimensional probability densities, particularly those lacking global regularity in the standard Euclidean setting. HRLMC leverages a Hessian manifold structure and mirror-descent techniques to model non-flat or constrained geometries, providing non-asymptotic convergence bounds in Wasserstein and Kullback–Leibler distances. This framework unifies Riemannian stochastic differential equations (SDEs), interior-point methods, and optimal transport theory, resolving difficulties in the analysis and computation of Riemannian discretizations by introducing relative smoothness, self-concordance, and explicit metric-dependent contraction rates (Zhang et al., 2020, Gatmiry et al., 2022).
1. Hessian Manifold Structure and Riemannian Metrics
Let be open and contractible, and let be a Legendre-type convex function (often called the "mirror potential" or barrier function). The Hessian manifold is endowed with the metric
where is positive definite for all . This construction allows dual coordinates , in which the metric becomes Euclidean and Riemannian quantities push forward via the mirror map. Important classes of include the Euclidean case and the log-barrier for polytope constraints, which yield (Zhang et al., 2020, Gatmiry et al., 2022).
2. Riemannian Langevin Diffusion: SDE Formulation
Let be a target density , . The Riemannian Langevin diffusion (RLD) on is the Itô SDE:
where
The invariant measure of this diffusion is by Fokker–Planck analysis. In coordinates the SDE simplifies, and the diffusion is natural Brownian motion on the manifold, with drift and diffusion (Zhang et al., 2020, Gatmiry et al., 2022).
3. Discrete HRLMC Scheme and Mirror Descent Analogy
The discretization leading to HRLMC is performed in the dual space and mapped back via the Legendre conjugate. With step-size and i.i.d. : HRLMC is the sampling analogue of mirror descent: setting yields the mirror descent step for with mirror map . The stochastic perturbation ensures the correct stationary distribution . In the Euclidean case, HRLMC recovers Unadjusted Langevin Algorithm (ULA) (Zhang et al., 2020). For constrained domains (e.g., polytopes), ULA in local charts uses the metric and steps: (Gatmiry et al., 2022).
4. Assumptions, Metric Properties, and Self-Concordance
Non-asymptotic convergence proofs require the following conditions:
- A1 (Self-concordance-like control): such that .
- A2 (Moment bound): .
- A3 (Relative strong convexity of ): with .
- A4 (Relative gradient Lipschitzness): such that .
- A5 (Commutator bound): .
Self-concordance is extended to second-order conditions: bounds on derivatives of the metric and its inverse, controlling the Jacobian growth and ensuring discretization errors remain manageable (Gatmiry et al., 2022).
5. Non-Asymptotic Convergence Bounds and Wasserstein Geometry
HRLMC delivers explicit, non-asymptotic contraction rates in the Wasserstein ground metric: With step-size and contraction factor , for all ,
where and explicit constants are supplied as functions of dimension, moment bounds, and regularity parameters. For constant step-size, iterates contract to an explicit Wasserstein ball around and remain inside, with radius proportional to and local geometry (Zhang et al., 2020).
For second-order self-concordant metrics, the convergence in Kullback–Leibler divergence is: where denotes KL divergence relative to and , are Lipschitz constants of and its Hessian (Gatmiry et al., 2022).
6. Specializations and Applications
Euclidean LMC: Setting yields , , recovers classical LMC with optimal Wasserstein bounds.
Constrained Sampling: For polytopes , using the log-barrier yields with self-concordant constants . HRLMC enables polynomial-time sampling of isoperimetric densities given gradient oracles and avoids global smoothness requirements.
Other Metrics: Adapting to the geometry of allows sampling from densities in non-Euclidean geometries, e.g., Dirichlet or gamma posteriors on the simplex (Zhang et al., 2020).
7. Key Analytical Tools and Proofs
HRLMC analysis relies on:
- Extended Baillon–Haddad Inequality: For relative smoothness, establishes cocoercivity bounds between gradients of and the mirror map.
- Itô’s Isometry and Minkowski Integral Inequality: Manage stochastic error terms and bias in diffusion discretization.
- Chung’s Lemma: Demonstrates contraction to a Wasserstein ball under vanishing step-sizes.
- Bismut–Elworthy–Li Formula: Controls higher-order terms in stochastic flows.
- Self-Concordance Theory: Manages metric derivatives and stability (Nesterov–Nemirovskii).
These tools together establish dimension-dependent rates, validate non-asymptotic guarantees, and ensure robustness to variance in local geometry (Zhang et al., 2020, Gatmiry et al., 2022).
HRLMC thus constitutes an advanced geometric Monte Carlo method tailored for high-dimensional, non-Euclidean, and constrained sampling problems, bridging stochastic analysis, convex optimization, and differential geometry. The rigorous contraction rates and metric-dependent bounds enable its application to non-smooth log-concave target distributions and provide theoretical foundations for sampling in complex domains.
Sponsored by Paperpile, the PDF & BibTeX manager trusted by top AI labs.
Get 30 days free