Training normalizing flows with computationally intensive target probability distributions (2308.13294v2)
Abstract: Machine learning techniques, in particular the so-called normalizing flows, are becoming increasingly popular in the context of Monte Carlo simulations as they can effectively approximate target probability distributions. In the case of lattice field theories (LFT) the target distribution is given by the exponential of the action. The common loss function's gradient estimator based on the "reparametrization trick" requires the calculation of the derivative of the action with respect to the fields. This can present a significant computational cost for complicated, non-local actions like e.g. fermionic action in QCD. In this contribution, we propose an estimator for normalizing flows based on the REINFORCE algorithm that avoids this issue. We apply it to two dimensional Schwinger model with Wilson fermions at criticality and show that it is up to ten times faster in terms of the wall-clock time as well as requiring up to $30\%$ less memory than the reparameterization trick estimator. It is also more numerically stable allowing for single precision calculations and the use of half-float tensor cores. We present an in-depth analysis of the origins of those improvements. We believe that these benefits will appear also outside the realm of the LFT, in each case where the target probability distribution is computationally intensive.
- Equation of state calculations by fast computing machines, The Journal of Chemical Physics 21 (1953) 1087–1092. doi:10.1063/1.1699114.
- W. K. Hastings, Monte Carlo sampling methods using Markov chains and their applications, Biometrika 57 (1970) 97–109. doi:10.1093/biomet/57.1.97.
- Flow-based generative models for markov chain monte carlo in lattice field theory, Phys. Rev. D 100 (2019) 034515.
- Solving statistical mechanics using variational autoregressive networks, Phys. Rev. Lett. 122 (2019) 080602.
- Asymptotically unbiased estimation of physical observables with neural samplers, Phys. Rev. E 101 (2020) 023304.
- Analysis of autocorrelation times in neural Markov chain Monte Carlo simulations, Phys. Rev. E 107 (2023) 015303. doi:10.1103/PhysRevE.107.015303. arXiv:2111.10189.
- Hierarchical autoregressive neural networks for statistical systems, Comput. Phys. Commun. 281 (2022) 108502. doi:10.1016/j.cpc.2022.108502. arXiv:2203.10989.
- Simulating first-order phase transition with hierarchical autoregressive networks, Phys. Rev. E 107 (2023) 054127. doi:10.1103/PhysRevE.107.054127. arXiv:2212.04955.
- D. P. Kingma, M. Welling, Auto-encoding variational bayes, 2nd International Conference on Learning Representations, ICLR 2014 - Conference Track Proceedings (2013). URL: https://arxiv.org/abs/1312.6114v11.
- Gradients should stay on path: better estimators of the reverse- and forward kl divergence for normalizing flows, Machine Learning: Science and Technology 3 (2022) 045006. doi:10.1088/2632-2153/ac9455.
- Flow-based sampling in the lattice schwinger model at criticality, Phys. Rev. D 106 (2022) 014514. doi:10.1103/PhysRevD.106.014514.
- J. S. Liu, Metropolized independent sampling with comparisons to rejection sampling and importance sampling, Statistics and Computing 6 (1996) 113–119.
- S. Kullback, R. A. Leibler, On Information and Sufficiency, The Annals of Mathematical Statistics 22 (1951) 79–86. doi:10.1214/aoms/1177729694.
- Estimation of thermodynamic observables in lattice field theories with deep generative models, Phys. Rev. Lett. 126 (2021) 032001. doi:10.1103/PhysRevLett.126.032001.
- Normalizing flows: An introduction and review of current methods, IEEE Transactions on Pattern Analysis and Machine Intelligence (2020) 1–1.
- A. Paszke, et al., Pytorch: An imperative style, high-performance deep learning library, in: Advances in Neural Information Processing Systems 32, Curran Associates, Inc., 2019, pp. 8024–8035.
- Automatic differentiation in machine learning: a survey, Journal of Machine Learning Research 18 (2018) 1–43. URL: http://jmlr.org/papers/v18/17-468.html.
- A. Mnih, D. J. Rezende, Variational inference for monte carlo objectives, in: Proceedings of the 33rd International Conference on International Conference on Machine Learning - Volume 48, ICML’16, JMLR.org, 2016, p. 2188–2196.
- Equivariant flow-based sampling for lattice gauge theory, Phys. Rev. Lett. 125 (2020) 121601. doi:10.1103/PhysRevLett.125.121601.
- Neural spline flows, 33rd Conference on Neural Information Processing Systems (NeurIPS 2019), Vancouver, Canada (2019). doi:10.48550/arxiv.1906.04032.
- Normalizing flows on tori and spheres, 37th International Conference on Machine Learning, ICML 2020 (2020) 8039–8048. doi:10.48550/arxiv.2002.02428.
- Gauge-equivariant flow models for sampling in lattice field theories with pseudofermions, Phys. Rev. D 106 (2022) 074506. URL: https://link.aps.org/doi/10.1103/PhysRevD.106.074506. doi:10.1103/PhysRevD.106.074506.
- Flow-based sampling for fermionic lattice field theories, Phys. Rev. D 104 (2021) 114507. doi:10.1103/PhysRevD.104.114507.
Collections
Sign up for free to add this paper to one or more collections.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.