Taming the Interacting Particle Langevin Algorithm: The Superlinear case (2403.19587v4)
Abstract: Recent advances in stochastic optimization have yielded the interacting particle Langevin algorithm (IPLA), which leverages the notion of interacting particle systems (IPS) to efficiently sample from approximate posterior densities. This becomes particularly crucial in relation to the framework of Expectation-Maximization (EM), where the E-step is computationally challenging or even intractable. Although prior research has focused on scenarios involving convex cases with gradients of log densities that grow at most linearly, our work extends this framework to include polynomial growth. Taming techniques are employed to produce an explicit discretization scheme that yields a new class of stable, under such non-linearities, algorithms which are called tamed interacting particle Langevin algorithms (tIPLA). We obtain non-asymptotic convergence error estimates in Wasserstein-2 distance for the new class under the best known rate.
- “Interacting Particle Langevin Algorithm for Maximum Marginal Likelihood Estimation”, 2023
- “The tamed unadjusted Langevin algorithm” In Stochastic Processes and their Applications 129.10, 2019, pp. 3638–3663
- “On Stochastic Gradient Langevin Dynamics with Dependent Data Streams: The Fully Nonconvex Case” In SIAM Journal on Mathematics of Data Science 3.3, 2021, pp. 959–986
- A.P. Dempster, N.M. Laird and D.B. Rubin “Maximum Likelihood from Incomplete Data via the EM Algorithm” In Journal of the Royal Statistical Society. Series B (Methodological) 39.1, 1977, pp. 1–38
- “High-dimensional Bayesian inference via the unadjusted Langevin algorithm” In Bernoulli 25.4A Bernoulli Society for Mathematical StatisticsProbability, 2019, pp. 2854–2882
- “Nonasymptotic convergence analysis for the unadjusted Langevin algorithm” In The Annals of Applied Probability 27.3 Institute of Mathematical Statistics, 2017, pp. 1551–1587
- Da Prato G “An Introduction to Infinite-Dimensional Analysis” Springer, 2006
- Martin Hutzenthaler, Arnulf Jentzen and Peter E. Kloeden “Strong and weak divergence in finite time of Euler’s method for stochastic differential equations with non-globally Lipschitz continuous coefficients” In Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences 467.2130 The Royal Society, 2010, pp. 1563–1576
- Martin Hutzenthaler, Arnulf Jentzen and Peter E. Kloeden “Strong convergence of an explicit numerical method for SDEs with nonglobally Lipschitz continuous coefficients” In The Annals of Applied Probability 22.4 Institute of Mathematical Statistics, 2012
- “On explicit approximations for Lévy driven SDEs with super-linear diffusion coefficients” In Electronic Journal of Probability 22.none Institute of Mathematical StatisticsBernoulli Society, 2017, pp. 1–19
- Juan Kuntz, Jen Ning Lim and Adam M. Johansen “Particle algorithms for maximum likelihood training of latent variable models” In Proceedings of The 26th International Conference on Artificial Intelligence and Statistics 206, Proceedings of Machine Learning Research PMLR, 2023, pp. 5134–5180
- “Polygonal Unadjusted Langevin Algorithms: Creating stable and efficient adaptive algorithms for neural networks”, 2021
- “Langevin dynamics based algorithm e-THϵitalic-ϵ\epsilonitalic_ϵO POULA for stochastic optimization problems with discontinuous stochastic gradient” In ArXiv abs/2210.13193, 2022
- “Non-asymptotic estimates for TUSLA algorithm for non-convex learning with applications to neural networks with ReLU activation function” In ArXiv abs/2107.08649, 2021
- “Taming Neural Networks with TUSLA: Nonconvex Learning via Adaptive Stochastic Gradient Langevin Algorithms” In SIAM Journal on Mathematics of Data Science 5.2, 2023, pp. 323–345
- Xuerong Mao “Stochastic Differential Equations and Applications” Elsevier, 2007
- Sotirios Sabanis “A note on tamed Euler approximations” In Electronic Communications in Probability 18.none Institute of Mathematical StatisticsBernoulli Society, 2013, pp. 1–10
- Sotirios Sabanis “Euler approximations with varying coefficients: The case of superlinearly growing diffusion coefficients” In The Annals of Applied Probability 26.4 Institute of Mathematical Statistics, 2016
- “Higher order Langevin Monte Carlo algorithm” In Electronic Journal of Statistics 13.2 Institute of Mathematical StatisticsBernoulli Society, 2019, pp. 3805–3850
- “On explicit order 1.5 approximations with varying coefficients: The case of super-linear diffusion coefficients” In Journal of Complexity 50 Elsevier BV, 2019, pp. 84–115
- De Bortoli V., Durmus A. and Pereyra M. al. “Efficient stochastic optimisation by unadjusted Langevin Monte Carlo” In Stat Comput 31.29, 2021