Learned Reference-based Diffusion Sampling for multi-modal distributions (2410.19449v3)
Abstract: Over the past few years, several approaches utilizing score-based diffusion have been proposed to sample from probability distributions, that is without having access to exact samples and relying solely on evaluations of unnormalized densities. The resulting samplers approximate the time-reversal of a noising diffusion process, bridging the target distribution to an easy-to-sample base distribution. In practice, the performance of these methods heavily depends on key hyperparameters that require ground truth samples to be accurately tuned. Our work aims to highlight and address this fundamental issue, focusing in particular on multi-modal distributions, which pose significant challenges for existing sampling methods. Building on existing approaches, we introduce Learned Reference-based Diffusion Sampler (LRDS), a methodology specifically designed to leverage prior knowledge on the location of the target modes in order to bypass the obstacle of hyperparameter tuning. LRDS proceeds in two steps by (i) learning a reference diffusion model on samples located in high-density space regions and tailored for multimodality, and (ii) using this reference model to foster the training of a diffusion-based sampler. We experimentally demonstrate that LRDS best exploits prior knowledge on the target distribution compared to competing algorithms on a variety of challenging distributions.
- Iterated denoising energy matching for sampling from boltzmann densities. In Proceedings of the 41st International Conference on Machine Learning, volume 235 of Proceedings of Machine Learning Research, pp. 760–786. PMLR, 21–27 Jul 2024. URL https://proceedings.mlr.press/v235/akhound-sadegh24a.html.
- Annealed flow transport Monte Carlo. In International Conference on Machine Learning, pp. 318–330. PMLR, 2021.
- A unified perspective on natural gradient variational inference with gaussian mixture models. Transactions on Machine Learning Research, 2023. ISSN 2835-8856. URL https://openreview.net/forum?id=tLBjsX4tjs.
- An optimal control perspective on diffusion-based generative modeling. Transactions on Machine Learning Research, 2023. URL https://openreview.net/forum?id=oYIjw37pTP.
- Beyond ELBOs: A large-scale evaluation of variational methods for sampling. In Proceedings of the 41st International Conference on Machine Learning, volume 235 of Proceedings of Machine Learning Research, pp. 4205–4229. PMLR, 21–27 Jul 2024. URL https://proceedings.mlr.press/v235/blessing24a.html.
- Handbook of Markov Chain Monte Carlo. Chapman and Hall/CRC, May 2011. ISBN 9780429138508. doi: 10.1201/b10905. URL http://dx.doi.org/10.1201/b10905.
- Time reversal of diffusion processes under a finite entropy condition. Annales de l’Institut Henri Poincaré, Probabilités et Statistiques, 59(4):1844 – 1881, 2023. doi: 10.1214/22-AIHP1320. URL https://doi.org/10.1214/22-AIHP1320.
- Lipschitz transport maps via the Föllmer flow. arXiv preprint arXiv:2309.03490, 2023.
- Arnak S Dalalyan. Theoretical guarantees for approximate sampling from smooth and log-concave densities. Journal of the Royal Statistical Society Series B: Statistical Methodology, 79(3):651–676, 2017.
- Target score matching. arXiv preprint arXiv:2402.08667, 2024.
- Sequential Monte Carlo samplers. Journal of the Royal Statistical Society Series B: Statistical Methodology, 68(3):411–436, 2006.
- Maximum likelihood from incomplete data via the em algorithm. Journal of the royal statistical society: series B (methodological), 39(1):1–22, 1977.
- Diffusion models beat GANs on image synthesis. In Advances in Neural Information Processing Systems, 2021. URL https://openreview.net/forum?id=AAWuCvzaVt.
- Score-based diffusion meets annealed importance sampling. Advances in Neural Information Processing Systems, 35:21482–21494, 2022.
- Hybrid monte carlo. Physics Letters B, 195(2):216–222, 1987. ISSN 0370-2693. doi: https://doi.org/10.1016/0370-2693(87)91197-X. URL https://www.sciencedirect.com/science/article/pii/037026938791197X.
- Quantitative bounds of convergence for geometrically ergodic Markov chain in the Wasserstein distance with application to the Metropolis Adjusted Langevin Algorithm. Statistics and Computing, 25:5–19, 2015.
- Nonasymptotic convergence analysis for the unadjusted Langevin algorithm. The Annals of Applied Probability, 27(3):1551 – 1587, 2017. doi: 10.1214/16-AAP1238. URL https://doi.org/10.1214/16-AAP1238.
- Adaptive monte carlo augmented with normalizing flows. Proceedings of the National Academy of Sciences, 119(10):e2109420119, 2022. doi: 10.1073/pnas.2109420119. URL https://www.pnas.org/doi/abs/10.1073/pnas.2109420119.
- Learning energy-based models by diffusion recovery likelihood. In International Conference on Learning Representations, 2021. URL https://openreview.net/forum?id=v_1Soh8QUNc.
- Mcmc variational inference via uncorrected hamiltonian annealing. Advances in Neural Information Processing Systems, 34:639–651, 2021.
- Langevin diffusion variational inference. In International Conference on Artificial Intelligence and Statistics, pp. 576–593. PMLR, 2023.
- On sampling with approximate transport maps. In Proceedings of the 40th International Conference on Machine Learning, volume 202 of Proceedings of Machine Learning Research, pp. 11698–11733. PMLR, 23–29 Jul 2023. URL https://proceedings.mlr.press/v202/grenioux23a.html.
- Stochastic localization via iterative posterior sampling. In Proceedings of the 41st International Conference on Machine Learning, volume 235 of Proceedings of Machine Learning Research, pp. 16337–16376. PMLR, 21–27 Jul 2024. URL https://proceedings.mlr.press/v235/grenioux24a.html.
- Denoising diffusion probabilistic models. Advances in neural information processing systems, 33:6840–6851, 2020.
- The no-u-turn sampler: Adaptively setting path lengths in hamiltonian monte carlo. Journal of Machine Learning Research, 15(47):1593–1623, 2014. URL http://jmlr.org/papers/v15/hoffman14a.html.
- PIPS: Path integral stochastic optimal control for path sampling in molecular dynamics, 2023. URL https://openreview.net/forum?id=TnIZfXSFJAh.
- Reverse Diffusion Monte Carlo. In The Twelfth International Conference on Learning Representations, 2024a. URL https://openreview.net/forum?id=kIPEyMSdFV.
- Faster sampling without isoperimetry via diffusion-based monte carlo. In Proceedings of Thirty Seventh Conference on Learning Theory, volume 247 of Proceedings of Machine Learning Research, pp. 2438–2493. PMLR, 30 Jun–03 Jul 2024b. URL https://proceedings.mlr.press/v247/huang24a.html.
- Estimation of non-normalized statistical models by score matching. Journal of Machine Learning Research, 6(4), 2005.
- Variational refinement for importance sampling using the forward Kullback-Leibler divergence. In Proceedings of the Thirty-Seventh Conference on Uncertainty in Artificial Intelligence, pp. 1819–1829. PMLR, December 2021. URL https://proceedings.mlr.press/v161/jerfel21a.html. ISSN: 2640-3498.
- Elucidating the design space of diffusion-based generative models. Advances in Neural Information Processing Systems, 35:26565–26577, 2022.
- Denoising MCMC for accelerating diffusion-based generative models. In Proceedings of the 40th International Conference on Machine Learning, volume 202 of Proceedings of Machine Learning Research, pp. 16955–16977. PMLR, 23–29 Jul 2023. URL https://proceedings.mlr.press/v202/kim23z.html.
- Werner Krauth. Statistical mechanics: algorithms and computations. OUP Oxford, 13, 2006.
- Handbook of Monte Carlo methods. Number 706 in Wiley series in probability and statistics. Wiley, Hoboken, 2011. ISBN 978-0-470-17793-8.
- Christian Léonard. Some properties of path measures. Séminaire de Probabilités XLVI, pp. 207–230, 2014.
- Monte Carlo strategies in scientific computing, volume 10. Springer, 2001.
- Continual repeated annealed flow transport monte carlo. In International Conference on Machine Learning, pp. 15196–15219. PMLR, 2022.
- Equation of State Calculations by Fast Computing Machines. The Journal of Chemical Physics, 21(6):1087–1092, 06 1953. ISSN 0021-9606. doi: 10.1063/1.1699114. URL https://doi.org/10.1063/1.1699114.
- Flow annealed importance sampling bootstrap. In The Eleventh International Conference on Learning Representations, 2023. URL https://openreview.net/forum?id=XCTVFJwS9LJ.
- Radford M Neal. Annealed importance sampling. Statistics and computing, 11:125–139, 2001.
- Radford M Neal. Mcmc using hamiltonian dynamics. arXiv preprint arXiv:1206.1901, 2012.
- Improved denoising diffusion probabilistic models. In International Conference on Machine Learning, pp. 8162–8171. PMLR, 2021.
- Boltzmann generators: Sampling equilibrium states of many-body systems with deep learning. Science, 365(6457):eaaw1147, September 2019. ISSN 0036-8075, 1095-9203. doi: 10.1126/science.aaw1147. URL https://www.science.org/doi/10.1126/science.aaw1147.
- Solving high-dimensional hamilton–jacobi–bellman pdes using neural networks: perspectives from the theory of controlled diffusions and measures on path space. Partial differential equations and applications, 2(4):48, 2021.
- Normalizing flows for probabilistic modeling and inference. Journal of Machine Learning Research, 22(57):1–64, 2021.
- Michele Pavon. On local entropy, stochastic control, and deep neural networks. IEEE Control Systems Letters, 7:437–441, 2022.
- Scikit-learn: Machine learning in python. Journal of machine learning research, 12(Oct):2825–2830, 2011.
- Particle denoising diffusion sampler. In Proceedings of the 41st International Conference on Machine Learning, volume 235 of Proceedings of Machine Learning Research, pp. 40688–40724. PMLR, 21–27 Jul 2024. URL https://proceedings.mlr.press/v235/phillips24a.html.
- A framework for adaptive MCMC targeting multimodal distributions. The Annals of Statistics, 48(5):2930–2952, October 2020. ISSN 0090-5364, 2168-8966. doi: 10.1214/19-AOS1916. URL https://projecteuclid.org/journals/annals-of-statistics/volume-48/issue-5/A-framework-for-adaptive-MCMC-targeting-multimodal-distributions/10.1214/19-AOS1916.full. Publisher: Institute of Mathematical Statistics.
- To smooth a cloud or to pin it down: Expressiveness guarantees and insights on score matching in denoising diffusion models. In The 40th Conference on Uncertainty in Artificial Intelligence, 2024. URL https://openreview.net/forum?id=BV2STKHKE6.
- Stochastic backpropagation and approximate inference in deep generative models. Proceedings of The 31st International Conference in Machine Learning, Beijing China, 32:1278–, 2014. URL http://jmlr.org/proceedings/papers/v32/rezende14.html%5Cnpapers3://publication/uuid/F2747569-7719-4EAC-A5A7-9ECA9D6A8FE6. arXiv: 1401.4082v3 ISBN: 9781634393973.
- Improved sampling via learned diffusions. In ICML Workshop on New Frontiers in Learning, Control, and Dynamical Systems, 2023. URL https://openreview.net/forum?id=uLgYD7ie0O.
- Exponential convergence of Langevin distributions and their discrete approximations. Bernoulli, 2(4):341–363, 1996. ISSN 13507265. URL http://www.jstor.org/stable/3318418.
- Should EBMs model the energy or the score?, 2021. URL https://openreview.net/forum?id=9AS-TF2jRNb.
- Deep unsupervised learning using nonequilibrium thermodynamics. In International conference on machine learning, pp. 2256–2265. PMLR, 2015.
- Generative modeling by estimating gradients of the data distribution. In Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc., 2019. URL https://proceedings.neurips.cc/paper_files/paper/2019/file/3001ef257407d5a371a96dcd947c7d93-Paper.pdf.
- Improved techniques for training score-based generative models. In Advances in Neural Information Processing Systems, volume 33, pp. 12438–12448. Curran Associates, Inc., 2020. URL https://proceedings.neurips.cc/paper_files/paper/2020/file/92c3b916311a5517d9290576e3ea37ad-Paper.pdf.
- Score-based generative modeling through stochastic differential equations. In The Ninth International Conference on Learning Representations, 2021. URL https://openreview.net/forum?id=PxTIG12RRHS.
- Free energy computations: A mathematical perspective. World Scientific, 2010.
- Replica Monte Carlo Simulation of Spin-Glasses. Physical Review Letters, 57(21):2607–2609, November 1986. doi: 10.1103/PhysRevLett.57.2607. URL https://link.aps.org/doi/10.1103/PhysRevLett.57.2607. Publisher: American Physical Society.
- Non-reversible parallel tempering: a scalable highly parallel mcmc scheme. Journal of the Royal Statistical Society Series B: Statistical Methodology, 84(2):321–350, 2022.
- Weight-preserving simulated tempering. Statistics and Computing, 30(1):27–41, 2020.
- Theoretical guarantees for sampling and inference in generative models with latent diffusions. In Conference on Learning Theory, pp. 3084–3114. PMLR, 2019.
- Denoising diffusion samplers. In The Eleventh International Conference on Learning Representations, 2023a. URL https://openreview.net/forum?id=8pvnfTAbu1f.
- Bayesian learning via neural Schrödinger–Föllmer flows. Statistics and Computing, 33(1):3, 2023b.
- Transport meets variational inference: Controlled monte carlo diffusions. In The Twelfth International Conference on Learning Representations, 2024. URL https://openreview.net/forum?id=PP1rudnxiW.
- Pascal Vincent. A connection between score matching and denoising autoencoders. Neural computation, 23(7):1661–1674, 2011.
- Graphical Models, Exponential Families, and Variational Inference. Foundations and Trends® in Machine Learning, 1:1–305, 2008. doi: 10.1561/2200000001. URL http://discovery.ucl.ac.uk/185880/.
- Sufficient Conditions for Torpid Mixing of Parallel and Simulated Tempering. Electronic Journal of Probability, 14(none):780 – 804, 2009. doi: 10.1214/EJP.v14-638. URL https://doi.org/10.1214/EJP.v14-638.
- Stochastic normalizing flows. Advances in Neural Information Processing Systems, 33:5933–5944, 2020.
- Path Integral Sampler: a stochastic control approach for sampling. In The Tenth International Conference on Learning Representations, 2022.
- Persistently trained, diffusion-assisted energy-based models. Stat, 12(1):e625, 2023. doi: https://doi.org/10.1002/sta4.625. URL https://onlinelibrary.wiley.com/doi/abs/10.1002/sta4.625.
- Learning energy-based models by cooperative diffusion recovery likelihood. In The Twelfth International Conference on Learning Representations, 2024. URL https://openreview.net/forum?id=AyzkDpuqcl.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.