Evidence Networks: simple losses for fast, amortized, neural Bayesian model comparison (2305.11241v2)
Abstract: Evidence Networks can enable Bayesian model comparison when state-of-the-art methods (e.g. nested sampling) fail and even when likelihoods or priors are intractable or unknown. Bayesian model comparison, i.e. the computation of Bayes factors or evidence ratios, can be cast as an optimization problem. Though the Bayesian interpretation of optimal classification is well-known, here we change perspective and present classes of loss functions that result in fast, amortized neural estimators that directly estimate convenient functions of the Bayes factor. This mitigates numerical inaccuracies associated with estimating individual model probabilities. We introduce the leaky parity-odd power (l-POP) transform, leading to the novel ``l-POP-Exponential'' loss function. We explore neural density estimation for data probability in different models, showing it to be less accurate and scalable than Evidence Networks. Multiple real-world and synthetic examples illustrate that Evidence Networks are explicitly independent of dimensionality of the parameter space and scale mildly with the complexity of the posterior probability density function. This simple yet powerful approach has broad implications for model inference tasks. As an application of Evidence Networks to real-world data we compute the Bayes factor for two models with gravitational lensing data of the Dark Energy Survey. We briefly discuss applications of our methods to other, related problems of model comparison and evaluation in implicit inference settings.
- Massive optimal data compression and density estimation for scalable, likelihood-free inference in cosmology. \mnras, 477(3):2874–2885, July 2018. doi: 10.1093/mnras/sty819.
- Fast likelihood-free cosmology with neural density estimators and active learning. \mnras, 488(3):4440–4458, September 2019. doi: 10.1093/mnras/stz1960.
- Dark energy survey year 3 results: Cosmology from cosmic shear and robustness to data calibration. Physical Review D, 105(2), jan 2022. doi: 10.1103/physrevd.105.023514. URL https://doi.org/10.1103%2Fphysrevd.105.023514.
- Relational inductive biases, deep learning, and graph networks, 2018.
- Mining for dark matter substructure: Inferring subhalo population properties from strong lenses with machine learning. The Astrophysical Journal, 886(1):49, Nov 2019. ISSN 1538-4357. doi: 10.3847/1538-4357/ab4c41. URL http://dx.doi.org/10.3847/1538-4357/ab4c41.
- Mining gold from implicit models to improve likelihood-free inference. Proceedings of the National Academy of Sciences, 117(10):5242–5249, 2020. ISSN 0027-8424. doi: 10.1073/pnas.1915980117. URL https://www.pnas.org/content/117/10/5242.
- Bayes and Empirical Bayes Methods for Data Analysis, Second Edition. Chapman & Hall/CRC Texts in Statistical Science. Taylor & Francis, 2010. ISBN 9781420057669. URL https://books.google.fr/books?id=484r1P5o0hQC.
- The frontier of simulation-based inference. Proceedings of the National Academy of Sciences, 2020. ISSN 0027-8424. doi: 10.1073/pnas.1912789117. URL https://www.pnas.org/content/early/2020/05/28/1912789117.
- PZFlow: normalizing flows for cosmology, with applications to forward modeling galaxy photometry. In American Astronomical Society Meeting Abstracts, volume 53 of American Astronomical Society Meeting Abstracts, pp. 230.01, June 2021.
- Dark energy survey year 3 results: cosmological constraints from the analysis of cosmic shear in harmonic space. Monthly Notices of the Royal Astronomical Society, 515(2):1942–1972, jul 2022. doi: 10.1093/mnras/stac1826. URL https://doi.org/10.1093%2Fmnras%2Fstac1826.
- Neural Spline Flows. arXiv e-prints, art. arXiv:1906.04032, June 2019. doi: 10.48550/arXiv.1906.04032.
- Bayes and the law. Annual Review of Statistics and Its Application, 3(1):51–77, 2016. doi: 10.1146/annurev-statistics-041715-033428. URL https://doi.org/10.1146/annurev-statistics-041715-033428. PMID: 27398389.
- Feroz, F. Calculation and applications of bayesian evidence in astrophysics and particle physics phenomenology. In 2013 IEEE 13th International Conference on Data Mining Workshops, pp. 8–15, 2013. doi: 10.1109/ICDMW.2013.21.
- MultiNest: an efficient and robust bayesian inference tool for cosmology and particle physics. Monthly Notices of the Royal Astronomical Society, 398(4):1601–1614, oct 2009. doi: 10.1111/j.1365-2966.2009.14548.x. URL https://doi.org/10.1111%2Fj.1365-2966.2009.14548.x.
- Bayesian Data Analysis, Third Edition. Chapman & Hall/CRC Texts in Statistical Science. Taylor & Francis, 2013. ISBN 9781439840955.
- Markov chain monte carlo methods for computing bayes factors. Journal of the American Statistical Association, 96(455):1122–1132, 2001. doi: 10.1198/016214501753208780. URL https://doi.org/10.1198/016214501753208780.
- polychord: next-generation nested sampling. Monthly Notices of the Royal Astronomical Society, 453(4):4385–4399, sep 2015. doi: 10.1093/mnras/stv1911. URL https://doi.org/10.1093%2Fmnras%2Fstv1911.
- Bayesian inflationary reconstructions from planck 2018 data. Physical Review D, 100(10), nov 2019. doi: 10.1103/physrevd.100.103511. URL https://doi.org/10.1103%2Fphysrevd.100.103511.
- Long short-term memory. Neural computation, 9(8):1735–1780, 1997.
- Batch normalization: Accelerating deep network training by reducing internal covariate shift, 2015.
- Nested sampling applied in bayesian room-acoustics decay analysis. The Journal of the Acoustical Society of America, 132:3251–62, 11 2012. doi: 10.1121/1.4754550.
- Jaynes, E. T. Probability theory: the logic of science. Cambridge University Press, 2003. ISBN 0521592712.
- Solving high-dimensional parameter inference: marginal posterior densities & Moment Networks. arXiv e-prints, art. arXiv:2011.05991, November 2020. doi: 10.48550/arXiv.2011.05991.
- Likelihood-free inference with neural compression of DES SV weak lensing map statistics, Nov 2020. ISSN 0035-8711. URL https://doi.org/10.1093/mnras/staa3594. staa3594, arXiv:2009.08459.
- Dark Energy Survey Year 3 results: Curved-sky weak lensing mass map reconstruction. \mnras, 505(3):4626–4645, August 2021. doi: 10.1093/mnras/stab1495.
- Jeffreys, H. The Theory of Probability. Oxford Classic Texts in the Physical Sciences. OUP Oxford, 1998. ISBN 9780191589676.
- Johnson, V. E. Revised standards for statistical evidence. Proceedings of the National Academy of Sciences, 110:19313 – 19317, 2013.
- Bayes factors. Journal of the American Statistical Association, 90(430):773–795, 1995. doi: 10.1080/01621459.1995.10476572. URL https://www.tandfonline.com/doi/abs/10.1080/01621459.1995.10476572.
- Using bayes factor hypothesis testing in neuroscience to establish evidence of absence. Nature neuroscience, 23(7):788–799, 2020.
- Adam: A method for stochastic optimization, 2017.
- Bayesian Evidence and Model Selection. arXiv e-prints, art. arXiv:1411.3013, November 2014. doi: 10.48550/arXiv.1411.3013.
- The sum of the masses of the Milky Way and M31: a likelihood-free inference approach. arXiv e-prints, art. arXiv:2010.08537, October 2020.
- Rectifier nonlinearities improve neural network acoustic models. In Proc. icml. Atlanta, Georgia, USA, 2013.
- On the design of loss functions for classification: theory, robustness to outliers, and savageboost. Advances in neural information processing systems, 21, 2008.
- Massimi, M. A Philosopher’s Look at the Dark Energy Survey: Reflections on the Use of the Bayes Factor in Cosmology. World Scientific, 2020. doi: 10.1142/9781786348364˙0025. URL https://www.worldscientific.com/doi/abs/10.1142/9781786348364_0025.
- Obtaining calibrated probabilities from boosting. In Proceedings of the Twenty-First Conference on Uncertainty in Artificial Intelligence, UAI’05, pp. 413–420, Arlington, Virginia, USA, 2005. AUAI Press. ISBN 0974903914.
- Fast ε𝜀\varepsilonitalic_ε-free inference of simulation models with bayesian conditional density estimation. Advances in Neural Information Processing Systems, pp. 1028–1036, 2016.
- Amortized Bayesian model comparison with evidential deep learning. arXiv e-prints, art. arXiv:2004.10629, April 2020. doi: 10.48550/arXiv.2004.10629.
- Dynamical mass inference of galaxy clusters with neural flows, 2020.
- Learning likelihood ratios with neural network classifiers, 2023.
- Dark energy survey year 3 results: Cosmology from cosmic shear and robustness to modeling uncertainty. Physical Review D, 105(2), jan 2022. doi: 10.1103/physrevd.105.023515. URL https://doi.org/10.1103%2Fphysrevd.105.023515.
- Bayesian model comparison for simulation-based inference. arXiv e-prints, art. arXiv:2207.04037, July 2022. doi: 10.48550/arXiv.2207.04037.
- Cosmic shear: Inference from forward models. Physical Review D, 100(2), Jul 2019. ISSN 2470-0029. doi: 10.1103/physrevd.100.023519. URL http://dx.doi.org/10.1103/PhysRevD.100.023519.
- Wakefield, J. Bayes factors for genome-wide association studies: comparison with p-values. Genetic Epidemiology: The Official Publication of the International Genetic Epidemiology Society, 33(1):79–86, 2009.
- Empirical Evaluation of Rectified Activations in Convolutional Network. arXiv e-prints, art. arXiv:1505.00853, May 2015. doi: 10.48550/arXiv.1505.00853.
- Yuen, K.-V. Recent developments of bayesian model class selection and applications in civil engineering. Structural Safety, 32(5):338–346, 2010.
- Transforming classifier scores into accurate multiclass probability estimates. In Proceedings of the Eighth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’02, pp. 694–699, New York, NY, USA, 2002. Association for Computing Machinery. ISBN 158113567X. doi: 10.1145/775047.775151. URL https://doi.org/10.1145/775047.775151.