Convergence of Kinetic Langevin Monte Carlo on Lie groups (2403.12012v2)
Abstract: Explicit, momentum-based dynamics for optimizing functions defined on Lie groups was recently constructed, based on techniques such as variational optimization and left trivialization. We appropriately add tractable noise to the optimization dynamics to turn it into a sampling dynamics, leveraging the advantageous feature that the trivialized momentum variable is Euclidean despite that the potential function lives on a manifold. We then propose a Lie-group MCMC sampler, by delicately discretizing the resulting kinetic-Langevin-type sampling dynamics. The Lie group structure is exactly preserved by this discretization. Exponential convergence with explicit convergence rate for both the continuous dynamics and the discrete sampler are then proved under $W_2$ distance. Only compactness of the Lie group and geodesically $L$-smoothness of the potential function are needed. To the best of our knowledge, this is the first convergence result for kinetic Langevin on curved spaces, and also the first quantitative result that requires no convexity or, at least not explicitly, any common relaxation such as isoperimetry.
- Foundations of mechanics. Foundations of Mechanics, 1978.
- ID Ado. The representation of lie algebras by matrices. Uspekhi Matematicheskikh Nauk, 2(6):159–173, 1947.
- From nesterov’s estimate sequence to riemannian acceleration. In Conference on Learning Theory, pages 84–118. PMLR, 2020.
- A continuous-time perspective for modeling acceleration in riemannian optimization. In International Conference on Artificial Intelligence and Statistics, pages 1297–1307. PMLR, 2020.
- Faster high-accuracy log-concave sampling via algorithmic warm starts. arXiv preprint arXiv:2302.10249, 2023.
- Irreversible langevin mcmc on lie groups. In Geometric Science of Information: 4th International Conference, GSI 2019, Toulouse, France, August 27–29, 2019, Proceedings 4, pages 171–179. Springer, 2019.
- Representations of compact Lie groups, volume 98. Springer Science & Business Media, 2013.
- Convergence of langevin mcmc in kl-divergence. In Algorithmic Learning Theory, pages 186–211. PMLR, 2018.
- Underdamped langevin mcmc: A non-asymptotic analysis. In Conference on learning theory, pages 300–323. PMLR, 2018.
- Efficient sampling on riemannian manifolds via langevin mcmc. Advances in Neural Information Processing Systems, 35:5995–6006, 2022.
- Sinho Chewi. Log-concave sampling. draft, 2024. URL https://chewisinho.github.io/main.pdf.
- Parseval networks: Improving robustness to adversarial examples. In International conference on machine learning, pages 854–863. PMLR, 2017.
- Arnak S Dalalyan. Theoretical guarantees for approximate sampling from smooth and log-concave densities. Journal of the Royal Statistical Society Series B: Statistical Methodology, 79(3):651–676, 2017.
- On sampling from a log-concave density using kinetic langevin diffusions. 2020.
- Analysis of langevin monte carlo via convex optimization. The Journal of Machine Learning Research, 20(1):2666–2711, 2019.
- EB Dynkin. Calculation of the coefficients in the campbell–hausdorff formula. DYNKIN, EB Selected Papers of EB Dynkin with Commentary. Ed. by YUSHKEVICH, AA, pages 31–35, 2000.
- Andreas Eberle. Reflection couplings and contraction rates for diffusions. Probability theory and related fields, 166:851–886, 2016.
- Couplings and quantitative contraction rates for langevin dynamics. 2019.
- Convergence of langevin monte carlo in chi-squared and rényi divergence. In International Conference on Artificial Intelligence and Statistics, pages 8151–8175. PMLR, 2022.
- Convergence of the riemannian langevin algorithm. arXiv preprint arXiv:2204.10818, 2022.
- A reduced parallel transport equation on lie groups with a left-invariant metric. In Geometric Science of Information: 5th International Conference, GSI 2021, Paris, France, July 21–23, 2021, Proceedings 5, pages 119–126. Springer, 2021.
- Elton P Hsu. Stochastic analysis on manifolds. Number 38. American Mathematical Soc., 2002.
- Momentum stiefel optimizer, with applications to suitably-orthogonal attention, and optimal transport. ICLR, 2023.
- Cheap orthogonal constraints in neural networks: A simple parametrization of the orthogonal and unitary group. In International Conference on Machine Learning, pages 3794–3803. PMLR, 2019.
- Sqrt(d) dimension dependence of langevin monte carlo. In International Conference on Learning Representations, 2021.
- Is there an analog of nesterov acceleration for gradient-based mcmc? Bernoulli, 2021.
- John Milnor. Curvatures of left invariant metrics on lie groups, 1976.
- Yurii Nesterov. Introductory lectures on convex optimization: A basic course, volume 87. Springer Science & Business Media, 2013.
- André Schlichting. Poincaré and log–sobolev inequalities for mixtures. Entropy, 21(1):89, 2019.
- The randomized midpoint method for log-concave sampling. Advances in Neural Information Processing Systems, 32, 2019.
- Variational optimization on lie groups, with examples of leading (generalized) eigenvalue problems. In International Conference on Artificial Intelligence and Statistics, pages 4269–4280. PMLR, 2020.
- Rapid convergence of the unadjusted langevin algorithm: Isoperimetry suffices. Advances in neural information processing systems, 32, 2019.
- Fast convergence of langevin dynamics on manifold: Geodesics meet log-sobolev. Advances in Neural Information Processing Systems, 33:18894–18904, 2020.
- Shing-Tung Yau. Non-existence of continuous convex functions on certain riemannian manifolds. Mathematische Annalen, 207:269–270, 1974.
- Markov chain monte carlo for gaussian: A linear control perspective. IEEE Control Systems Letters, 2023.
- Towards riemannian accelerated gradient methods. COLT, 2018.
- Improved discretization analysis for underdamped langevin monte carlo. In The Thirty Sixth Annual Conference on Learning Theory, pages 36–71. PMLR, 2023.