Calibrating Neural Simulation-Based Inference with Differentiable Coverage Probability (2310.13402v1)
Abstract: Bayesian inference allows expressing the uncertainty of posterior belief under a probabilistic model given prior information and the likelihood of the evidence. Predominantly, the likelihood function is only implicitly established by a simulator posing the need for simulation-based inference (SBI). However, the existing algorithms can yield overconfident posteriors (Hermans et al., 2022) defeating the whole purpose of credibility if the uncertainty quantification is inaccurate. We propose to include a calibration term directly into the training objective of the neural model in selected amortized SBI techniques. By introducing a relaxation of the classical formulation of calibration error we enable end-to-end backpropagation. The proposed method is not tied to any particular neural model and brings moderate computational overhead compared to the profits it introduces. It is directly applicable to existing computational pipelines allowing reliable black-box posterior inference. We empirically show on six benchmark problems that the proposed method achieves competitive or better results in terms of coverage and expected posterior density than the previously existing approaches.
- Observation of a new particle in the search for the Standard Model Higgs boson with the ATLAS detector at the LHC. Physics Letters B, 716(1):1–29, 2012.
- Efficient sensor placement optimization using gradient descent and probabilistic coverage. Sensors, 14(8):15525–15552, 2014.
- Adaptive approximate Bayesian computation. Biometrika, 96(4):983–990, 2009.
- Estimating or propagating gradients through stochastic neurons for conditional computation. arXiv preprint arXiv:1308.3432, 2013.
- V. W. Berger and Y. Zhou. Kolmogorov–Smirnov Test: Overview. John Wiley & Sons, Ltd, 2014.
- Fast differentiable sorting and ranking. In International Conference on Machine Learning, pages 950–959. PMLR, 2020.
- Handbook of Markov chain Monte Carlo. CRC press, 2011.
- Some aspects of ‘characterization’ problems in geomechanics. In International Journal of Rock Mechanics and Mining Sciences & Geomechanics Abstracts, volume 18, pages 487–503. Elsevier, 1981.
- The frontier of simulation-based inference. Proceedings of the National Academy of Sciences, 117(48):30055–30062, 2020.
- A Modern Introduction to Probability and Statistics: Understanding why and how, volume 488, page 219. Springer, 2005.
- Towards reliable simulation-based inference with balanced neural ratio estimation. In A. H. Oh, A. Agarwal, D. Belgrave, and K. Cho, editors, Advances in Neural Information Processing Systems, 2022.
- Balancing simulation-based inference for conservative posteriors. arXiv preprint arXiv:2304.10978, 2023.
- Calibrated predictive distributions via diagnostics for conditional coverage. arXiv preprint arXiv:2205.14568, 2022.
- Monte Carlo methods of inference for implicit statistical models. Journal of the Royal Statistical Society. Series B (Methodological), 46(2):193–227, 1984.
- On contrastive learning for likelihood-free inference. In International Conference on Machine Learning, pages 2771–2781. PMLR, 2020.
- Score modeling for simulation-based inference. In NeurIPS 2022 Workshop on Score-Based Methods, 2022.
- J. Geweke. Bayesian inference in econometric models using Monte Carlo integration. Econometrica: Journal of the Econometric Society, pages 1317–1339, 1989.
- Automatic posterior transformation for likelihood-free inference. In International Conference on Machine Learning, pages 2404–2414. PMLR, 2019.
- Likelihood-free MCMC with amortized approximate ratio estimators. In International conference on machine learning, pages 4239–4248. PMLR, 2020.
- A crisis in simulation-based inference? Beware, your posterior approximations can be unfaithful. Transactions on Machine Learning Research, 2022.
- J. L. Horowitz. Ill-posed inverse problems in economics. Annu. Rev. Econ., 6(1):21–51, 2014.
- R. J. Hyndman. Computing and graphing highest density regions. The American Statistician, 50(2):120–126, 1996.
- E. T. Jaynes. Prior information and ambiguity in inverse problems. Inverse problems, 14:151–166, 1984.
- D. P. Kingma and M. Welling. Auto-encoding variational Bayes. arXiv preprint arXiv:1312.6114, 2013.
- Self-normalizing neural networks. Advances in neural information processing systems, 30, 2017.
- Normalizing flows: An introduction and review of current methods. IEEE transactions on pattern analysis and machine intelligence, 43(11):3964–3979, 2020.
- Sampling-based accuracy testing of posterior estimators for general inference. In A. Krause, E. Brunskill, K. Cho, B. Engelhardt, S. Sabato, and J. Scarlett, editors, Proceedings of the 40th International Conference on Machine Learning, volume 202 of Proceedings of Machine Learning Research, pages 19256–19273. PMLR, 23–29 Jul 2023.
- Validation Diagnostics for SBI algorithms based on Normalizing Flows. In NeurIPS 2022 Machine Learning and the Physical Sciences Workshop, pages 1–7, New Orleans, United States, Nov. 2022.
- Calculation of susceptibility through multiple orientation sampling (COSMOS): a method for conditioning the inverse problem from measured magnetic field map to susceptibility source image in MRI. Magnetic Resonance in Medicine: An Official Journal of the International Society for Magnetic Resonance in Medicine, 61(1):196–204, 2009.
- Flexible statistical inference for mechanistic models of neural dynamics. In Proceedings of the 31st International Conference on Neural Information Processing Systems, pages 1289–1299, 2017.
- Benchmarking simulation-based inference. In International Conference on Artificial Intelligence and Statistics, pages 343–351. PMLR, 2021.
- Contrastive neural ratio estimation. Advances in Neural Information Processing Systems, 35:3262–3278, 2022.
- Recent progress on generative adversarial networks (GANs): A survey. IEEE access, 7:36322–36333, 2019.
- G. Papamakarios and I. Murray. Fast ε𝜀\varepsilonitalic_ε-free inference of simulation models with Bayesian conditional density estimation. Advances in neural information processing systems, 29, 2016.
- Sequential neural likelihood: Fast likelihood-free inference with autoregressive flows. In The 22nd International Conference on Artificial Intelligence and Statistics, pages 837–848. PMLR, 2019.
- Normalizing flows for probabilistic modeling and inference. The Journal of Machine Learning Research, 22(1):2617–2680, 2021.
- The emergence of spectral universality in deep networks. In International Conference on Artificial Intelligence and Statistics, pages 1924–1932. PMLR, 2018.
- Population growth of human Y chromosomes: a study of Y chromosome microsatellites. Molecular biology and evolution, 16(12):1791–1798, 1999.
- J. P. Quirk and R. Saposnik. Admissibility and measurable utility functions. The Review of Economic Studies, 29(2):140–146, 1962.
- D. B. Rubin. Bayesianly justifiable and relevant frequency calculations for the applied statistician. The Annals of Statistics, 12(4):1151 – 1172, 1984.
- Sequential neural score estimation: Likelihood-free inference with conditional score based diffusion models. arXiv preprint arXiv:2210.04872, 2022.
- Sequential Monte Carlo without likelihoods. Proceedings of the National Academy of Sciences, 104(6):1760–1765, 2007.
- Handbook of approximate Bayesian computation. CRC Press, 2018.
- Approximate Bayesian computation. PLOS Computational Biology, 9(1):1–10, 01 2013.
- Validating Bayesian inference algorithms with simulation-based calibration. arXiv preprint arXiv:1804.06788, 2018.
- A. Tarantola. Inverse Problem Theory and Methods for Model Parameter Estimation. Society for Industrial and Applied Mathematics, 2005.
- J. M. Tomczak. Deep generative modeling. Springer, 2022.
- T. Toni and M. P. H. Stumpf. Simulation-based model selection for dynamical systems in systems and population biology. Bioinformatics, 26(1):104–110, 10 2009.
- F. Topsøe. On the Glivenko-Cantelli theorem. Zeitschrift für Wahrscheinlichkeitstheorie und verwandte Gebiete, 14(3):239–250, 1970.
- Gradients should stay on path: better estimators of the reverse- and forward KL divergence for normalizing flows. Machine Learning: Science and Technology, 3(4):045006, oct 2022.
- V. Volodina and P. Challenor. The importance of uncertainty quantification in model reproducibility. Philosophical Transactions of the Royal Society A, 379(2197):20200071, 2021.
- Diffusion models: A comprehensive survey of methods and applications. ACM Comput. Surv., sep 2023. Just Accepted.
- Diagnostics for conditional density models and Bayesian inference algorithms. In Uncertainty in Artificial Intelligence, pages 1830–1840. PMLR, 2021.