Papers
Topics
Authors
Recent
Detailed Answer
Quick Answer
Concise responses based on abstracts only
Detailed Answer
Well-researched responses based on abstracts and relevant paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses
Gemini 2.5 Flash
Gemini 2.5 Flash 64 tok/s
Gemini 2.5 Pro 50 tok/s Pro
GPT-5 Medium 30 tok/s Pro
GPT-5 High 35 tok/s Pro
GPT-4o 77 tok/s Pro
Kimi K2 174 tok/s Pro
GPT OSS 120B 457 tok/s Pro
Claude Sonnet 4 37 tok/s Pro
2000 character limit reached

Training normalizing flows with computationally intensive target probability distributions (2308.13294v2)

Published 25 Aug 2023 in cs.LG, cond-mat.stat-mech, and hep-lat

Abstract: Machine learning techniques, in particular the so-called normalizing flows, are becoming increasingly popular in the context of Monte Carlo simulations as they can effectively approximate target probability distributions. In the case of lattice field theories (LFT) the target distribution is given by the exponential of the action. The common loss function's gradient estimator based on the "reparametrization trick" requires the calculation of the derivative of the action with respect to the fields. This can present a significant computational cost for complicated, non-local actions like e.g. fermionic action in QCD. In this contribution, we propose an estimator for normalizing flows based on the REINFORCE algorithm that avoids this issue. We apply it to two dimensional Schwinger model with Wilson fermions at criticality and show that it is up to ten times faster in terms of the wall-clock time as well as requiring up to $30\%$ less memory than the reparameterization trick estimator. It is also more numerically stable allowing for single precision calculations and the use of half-float tensor cores. We present an in-depth analysis of the origins of those improvements. We believe that these benefits will appear also outside the realm of the LFT, in each case where the target probability distribution is computationally intensive.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (23)
  1. Equation of state calculations by fast computing machines, The Journal of Chemical Physics 21 (1953) 1087–1092. doi:10.1063/1.1699114.
  2. W. K. Hastings, Monte Carlo sampling methods using Markov chains and their applications, Biometrika 57 (1970) 97–109. doi:10.1093/biomet/57.1.97.
  3. Flow-based generative models for markov chain monte carlo in lattice field theory, Phys. Rev. D 100 (2019) 034515.
  4. Solving statistical mechanics using variational autoregressive networks, Phys. Rev. Lett. 122 (2019) 080602.
  5. Asymptotically unbiased estimation of physical observables with neural samplers, Phys. Rev. E 101 (2020) 023304.
  6. Analysis of autocorrelation times in neural Markov chain Monte Carlo simulations, Phys. Rev. E 107 (2023) 015303. doi:10.1103/PhysRevE.107.015303. arXiv:2111.10189.
  7. Hierarchical autoregressive neural networks for statistical systems, Comput. Phys. Commun. 281 (2022) 108502. doi:10.1016/j.cpc.2022.108502. arXiv:2203.10989.
  8. Simulating first-order phase transition with hierarchical autoregressive networks, Phys. Rev. E 107 (2023) 054127. doi:10.1103/PhysRevE.107.054127. arXiv:2212.04955.
  9. D. P. Kingma, M. Welling, Auto-encoding variational bayes, 2nd International Conference on Learning Representations, ICLR 2014 - Conference Track Proceedings (2013). URL: https://arxiv.org/abs/1312.6114v11.
  10. Gradients should stay on path: better estimators of the reverse- and forward kl divergence for normalizing flows, Machine Learning: Science and Technology 3 (2022) 045006. doi:10.1088/2632-2153/ac9455.
  11. Flow-based sampling in the lattice schwinger model at criticality, Phys. Rev. D 106 (2022) 014514. doi:10.1103/PhysRevD.106.014514.
  12. J. S. Liu, Metropolized independent sampling with comparisons to rejection sampling and importance sampling, Statistics and Computing 6 (1996) 113–119.
  13. S. Kullback, R. A. Leibler, On Information and Sufficiency, The Annals of Mathematical Statistics 22 (1951) 79–86. doi:10.1214/aoms/1177729694.
  14. Estimation of thermodynamic observables in lattice field theories with deep generative models, Phys. Rev. Lett. 126 (2021) 032001. doi:10.1103/PhysRevLett.126.032001.
  15. Normalizing flows: An introduction and review of current methods, IEEE Transactions on Pattern Analysis and Machine Intelligence (2020) 1–1.
  16. A. Paszke, et al., Pytorch: An imperative style, high-performance deep learning library, in: Advances in Neural Information Processing Systems 32, Curran Associates, Inc., 2019, pp. 8024–8035.
  17. Automatic differentiation in machine learning: a survey, Journal of Machine Learning Research 18 (2018) 1–43. URL: http://jmlr.org/papers/v18/17-468.html.
  18. A. Mnih, D. J. Rezende, Variational inference for monte carlo objectives, in: Proceedings of the 33rd International Conference on International Conference on Machine Learning - Volume 48, ICML’16, JMLR.org, 2016, p. 2188–2196.
  19. Equivariant flow-based sampling for lattice gauge theory, Phys. Rev. Lett. 125 (2020) 121601. doi:10.1103/PhysRevLett.125.121601.
  20. Neural spline flows, 33rd Conference on Neural Information Processing Systems (NeurIPS 2019), Vancouver, Canada (2019). doi:10.48550/arxiv.1906.04032.
  21. Normalizing flows on tori and spheres, 37th International Conference on Machine Learning, ICML 2020 (2020) 8039–8048. doi:10.48550/arxiv.2002.02428.
  22. Gauge-equivariant flow models for sampling in lattice field theories with pseudofermions, Phys. Rev. D 106 (2022) 074506. URL: https://link.aps.org/doi/10.1103/PhysRevD.106.074506. doi:10.1103/PhysRevD.106.074506.
  23. Flow-based sampling for fermionic lattice field theories, Phys. Rev. D 104 (2021) 114507. doi:10.1103/PhysRevD.104.114507.
Citations (4)
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-Up Questions

We haven't generated follow-up questions for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com