Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash 94 tok/s
Gemini 2.5 Pro 39 tok/s Pro
GPT-5 Medium 20 tok/s
GPT-5 High 16 tok/s Pro
GPT-4o 99 tok/s
GPT OSS 120B 476 tok/s Pro
Kimi K2 214 tok/s Pro
2000 character limit reached

Projected Langevin Monte Carlo algorithms in non-convex and super-linear setting (2312.17077v4)

Published 28 Dec 2023 in math.NA, cs.NA, and math.PR

Abstract: It is of significant interest in many applications to sample from a high-dimensional target distribution $\pi$ with the density $\pi(\text{d} x) \propto e{-U(x)} (\text{d} x) $, based on the temporal discretization of the Langevin stochastic differential equations (SDEs). In this paper, we propose an explicit projected Langevin Monte Carlo (PLMC) algorithm with non-convex potential $U$ and super-linear gradient of $U$ and investigate the non-asymptotic analysis of its sampling error in total variation distance. Equipped with time-independent regularity estimates for the associated Kolmogorov equation, we derive the non-asymptotic bounds on the total variation distance between the target distribution of the Langevin SDEs and the law induced by the PLMC scheme with order $\mathcal{O}(d{\max{3\gamma/2 , 2\gamma-1 }} h |\ln h|)$, where $d$ is the dimension of the target distribution and $\gamma \geq 1$ characterizes the growth of the gradient of $U$. In addition, if the gradient of $U$ is globally Lipschitz continuous, an improved convergence order of $\mathcal{O}(d{3/2} h)$ for the classical Langevin Monte Carlo (LMC) scheme is derived with a refinement of the proof based on Malliavin calculus techniques. To achieve a given precision $\epsilon$, the smallest number of iterations of the PLMC algorithm is proved to be of order ${\mathcal{O}}\big(\tfrac{d{\max{3\gamma/2 , 2\gamma-1 }}}{\epsilon} \ \cdot \ln (\tfrac{d}{\epsilon}) \cdot \ln (\tfrac{1}{\epsilon}) \big)$. In particular, the classical Langevin Monte Carlo (LMC) scheme with the non-convex potential $U$ and the globally Lipschitz gradient of $U$ can be guaranteed by order ${\mathcal{O}}\big(\tfrac{d{3/2}}{\epsilon} \cdot \ln (\tfrac{1}{\epsilon}) \big)$. Numerical experiments are provided to confirm the theoretical findings.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (28)
  1. Stochastic C-stability and B-consistency of explicit and implicit Euler-type schemes. Journal of Scientific Computing, 67(3):955–987, 2016.
  2. Stochastic C-stability and B-consistency of explicit and implicit Milstein-type schemes. Journal of Scientific Computing, 70(3):1042–1077, 2017.
  3. C.-E. Bréhier. Approximation of the invariant measure with an Euler scheme for stochastic PDEs driven by space-time white noise. Potential Analysis, 40(1):1–40, 2014.
  4. The tamed unadjusted Langevin algorithm. Stochastic Processes and their Applications, 129(10):3638–3663, 2019.
  5. S. Cerrai. Second order PDE’s in finite and infinite dimension: a probabilistic approach. Springer, 2001.
  6. Sharp convergence rates for Langevin dynamics in the nonconvex setting. arXiv preprint arXiv:1805.01648, 2018.
  7. G. Da Prato. An introduction to infinite-dimensional analysis. Springer Science & Business Media, 2006.
  8. A. S. Dalalyan. Theoretical guarantees for approximate sampling from smooth and log-concave densities. Journal of the Royal Statistical Society Series B: Statistical Methodology, 79(3):651–676, 2017.
  9. A. Durmus and É. Moulines. Nonasymptotic convergence analysis for the unadjusted Langevin algorithm. The Annals of Applied Probability, 27(3):1551–1587, 2017.
  10. A. Durmus and É. Moulines. High-dimensional Bayesian inference via the unadjusted Langevin algorithm. Bernoulli, 25(4A):2854–2882, 2019.
  11. Formulae for the derivatives of heat semigroups. Journal of Functional Analysis, 125(1):252–286, 1994.
  12. B. Goldys. Exponential Ergodicity for Stochastic Reaction–Diffusion Equations. Stochastic Partial Differential Equations and Applications-VII, page 115, 2005.
  13. Strong and weak divergence in finite time of Euler’s method for stochastic differential equations with non-globally Lipschitz continuous coefficients. Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences, 467(2130):1563–1576, 2011.
  14. Strong convergence of an explicit numerical method for SDEs with nonglobally Lipschitz continuous coefficients. The Annals of Applied Probability, 22(4):1611–1641, 2012.
  15. Sqrt (d) Dimension Dependence of Langevin Monte Carlo. In The International Conference on Learning Representations, 2022.
  16. Explicit numerical approximations for stochastic differential equations in finite and infinite horizons: truncation methods, convergence in pth moment and stability. IMA Journal of Numerical Analysis, 39(2):847–892, 2019.
  17. Unadjusted Langevin Algorithms for SDEs with Hölder Drift. arXiv preprint arXiv:2310.00232, 2023.
  18. Nonasymptotic bounds for sampling algorithms without log-concavity. The Annals of Applied Probability, 30(4):1534–1581, 2020.
  19. Ergodicity for SDEs and approximations: locally Lipschitz vector fields and degenerate noise. Stochastic processes and their applications, 101(2):185–232, 2002.
  20. Non-asymptotic convergence bounds for modified tamed unadjusted Langevin algorithm in non-convex setting. arXiv preprint arXiv:2207.02600, 2022.
  21. G. Pages and F. Panloup. Unadjusted Langevin algorithm with multiplicative noise: Total variation and Wasserstein bounds. The Annals of Applied Probability, 33(1):726–779, 2023.
  22. Linear implicit approximations of invariant measures of semi-linear SDEs with non-globally Lipschitz coefficients. arXiv preprint arXiv:2308.12886, 2023.
  23. S. Sabanis. Euler approximations with varying coefficients: the case of superlinearly growing diffusion coefficients. The Annals of Applied Probability, 26(4):2083–2105, 2016.
  24. Ł. Szpruch and X. Zhāng. V-integrability, asymptotic stability and comparison property of explicit numerical schemes for non-linear SDEs. Mathematics of Computation, 87(310):755–783, 2018.
  25. X. Wang and S. Gan. The tamed Milstein method for commutative stochastic differential equations with non-globally Lipschitz continuous coefficients. Journal of Difference Equations and Applications, 19(3):466–490, 2013.
  26. Mean-square convergence rates of stochastic theta methods for SDEs under a coupled monotonicity condition. BIT Numerical Mathematics, 60(3):759–790, 2020.
  27. Weak error analysis for strong approximation schemes of SDEs with super-linear coefficients. IMA Journal of Numerical Analysis, page drad083, 11 2023.
  28. Second-order numerical methods of weak convergence for SDEs with super-linear coefficients. preprint, 2023.
Citations (1)
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

We haven't generated a summary for this paper yet.

Ai Generate Text Spark Streamline Icon: https://streamlinehq.com

Paper Prompts

Sign up for free to create and run prompts on this paper using GPT-5.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube