All-in-one simulation-based inference (2404.09636v3)
Abstract: Amortized Bayesian inference trains neural networks to solve stochastic inference problems using model simulations, thereby making it possible to rapidly perform Bayesian inference for any newly observed data. However, current simulation-based amortized inference methods are simulation-hungry and inflexible: They require the specification of a fixed parametric prior, simulator, and inference tasks ahead of time. Here, we present a new amortized inference method -- the Simformer -- which overcomes these limitations. By training a probabilistic diffusion model with transformer architectures, the Simformer outperforms current state-of-the-art amortized inference approaches on benchmark tasks and is substantially more flexible: It can be applied to models with function-valued parameters, it can handle inference scenarios with missing or unstructured data, and it can sample arbitrary conditionals of the joint distribution of parameters and data, including both posterior and likelihood. We showcase the performance and flexibility of the Simformer on simulators from ecology, epidemiology, and neuroscience, and demonstrate that it opens up new possibilities and application domains for amortized Bayesian inference on simulation-based models.
- Stochastic interpolants: A unifying framework for flows and diffusions. arXiv preprint arXiv:2303.08797, 2023.
- Anderson, B. D. Reverse-time diffusion equation models. Stochastic Processes and their Applications, 12(3):313–326, 1982.
- Universal guidance for diffusion models. In 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 843–852, 2023. doi: 10.1109/CVPRW59228.2023.00091.
- Adaptive approximate bayesian computation. Biometrika, 2009.
- JAX: composable transformations of Python+NumPy programs, 2018.
- A Likelihood-Free inference framework for population genetic data using exchangeable neural networks. Adv Neural Inf Process Syst, 31:8594–8605, December 2018.
- A time-dependent sir model for covid-19 with undetectable infected persons. IEEE Transactions on Network Science and Engineering, 7(4):3279–3294, October 2020. ISSN 2334-329X. doi: 10.1109/tnse.2020.3024723.
- Improving diffusion models for inverse problems using manifold constraints. Advances in Neural Information Processing Systems, 35:25683–25696, 2022.
- Diffusion posterior sampling for general noisy inverse problems. In The Eleventh International Conference on Learning Representations, 2023.
- The frontier of simulation-based inference. Proceedings of the National Academy of Sciences, 117(48):30055–30062, 2020.
- Real-time gravitational wave science with neural posterior estimation. Phys. Rev. Lett., 127:241103, Dec 2021. doi: 10.1103/PhysRevLett.127.241103.
- Truncated proposals for scalable and hassle-free simulation-based inference. In Oh, A. H., Agarwal, A., Belgrave, D., and Cho, K. (eds.), Advances in Neural Information Processing Systems, 2022a.
- Energy-efficient network activity from disparate circuit parameters. Proceedings of the National Academy of Sciences, 119(44):e2207632119, 2022b.
- Diffusion models beat gans on image synthesis. In Ranzato, M., Beygelzimer, A., Dauphin, Y., Liang, P., and Vaughan, J. W. (eds.), Advances in Neural Information Processing Systems, volume 34, pp. 8780–8794. Curran Associates, Inc., 2021.
- On contrastive learning for likelihood-free inference. In International Conference on Machine Learning, pp. 2771–2781. PMLR, 2020.
- Approximate bayesian computation with path signatures. arXiv preprint arXiv:2106.12555, 2021.
- Sensitivity-aware amortized bayesian inference. arXiv preprint arXiv:2310.11122, 2023.
- Compositional score modeling for simulation-based inference. In International Conference on Machine Learning, pp. 11098–11116. PMLR, 2023.
- Made: Masked autoencoder for distribution estimation. In Bach, F. and Blei, D. (eds.), Proceedings of the 32nd International Conference on Machine Learning, volume 37 of Proceedings of Machine Learning Research, pp. 881–889, Lille, France, 07–09 Jul 2015. PMLR.
- Variational methods for simulation-based inference. In International Conference on Learning Representations, 2022.
- Training deep neural density estimators to identify mechanistic models of neural dynamics. Elife, 9:e56261, 2020.
- Seeds: Exponential sde solvers for fast high-quality sampling from diffusion models, 2023.
- Automatic posterior transformation for likelihood-free inference. In International Conference on Machine Learning, pp. 2404–2414. PMLR, 2019.
- Rethinking the objectives of vector-quantized tokenizers for image synthesis. arXiv preprint arXiv:2212.03185, 2022.
- Flexible diffusion modeling of long videos. Advances in Neural Information Processing Systems, 35:27953–27965, 2022.
- Diffit: Diffusion vision transformers for image generation, 2023.
- Likelihood-free mcmc with amortized approximate ratio estimators. In International Conference on Machine Learning, pp. 4239–4248. PMLR, 2020.
- Classifier-free diffusion guidance. In NeurIPS 2021 Workshop on Deep Generative Models and Downstream Applications, 2021.
- A quantitative description of membrane current and its application to conduction and excitation in nerve. J Physiol, 117(4):500–544, Aug 1952. doi: 10.1113/jphysiol.1952.sp004764.
- Estimation of non-normalized statistical models by score matching. Journal of Machine Learning Research, 6(4), 2005.
- Variational autoencoder with arbitrary conditioning. In International Conference on Learning Representations, 2019.
- Robust compressed sensing mri with deep generative priors. In Ranzato, M., Beygelzimer, A., Dauphin, Y., Liang, P., and Vaughan, J. W. (eds.), Advances in Neural Information Processing Systems, volume 34, pp. 14938–14954. Curran Associates, Inc., 2021.
- Sparse is enough in scaling transformers, 2021.
- A contribution to the mathematical theory of epidemics. Proceedings of the royal society of london. Series A, Containing papers of a mathematical and physical character, 115(772):700–721, 1927.
- Inference compilation and universal probabilistic programming. In Artificial Intelligence and Statistics, pp. 1338–1348. PMLR, 2017.
- Acflow: Flow models for arbitrary conditional likelihoods. In International Conference on Machine Learning, pp. 5831–5841. PMLR, 2020.
- A survey of transformers. AI Open, 2022.
- Lotka, A. J. Elements of physical biology. Williams & Wilkins, 1925.
- Flexible statistical inference for mechanistic models of neural dynamics. Advances in neural information processing systems, 30, 2017.
- Benchmarking simulation-based inference. In International Conference on Artificial Intelligence and Statistics, pp. 343–351. PMLR, 2021.
- Repaint: Inpainting using denoising diffusion probabilistic models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 11461–11471, 2022.
- Handbook of graphical models. CRC Press, 2018.
- End-to-end meta-bayesian optimisation with transformer neural processes. arXiv preprint arXiv:2305.15930, 2023.
- Simulation-based bayesian inference for robotic grasping. In IROS 2022 Workshop Probabilistic Robotics in the Age of Deep Learning, 2022.
- Simulation-based inference of surface accumulation and basal melt rates of an antarctic ice shelf from isochronal layers, 2023.
- Transformers can do bayesian inference, 2023.
- Transformer neural processes: Uncertainty-aware meta learning via sequence modeling. arXiv preprint arXiv:2207.04179, 2022a.
- Transformer neural processes: Uncertainty-aware meta learning via sequence modeling. arXiv preprint arXiv:2207.04179, 2022b.
- Fast ε𝜀\varepsilonitalic_ε-free inference of simulation models with bayesian conditional density estimation. Advances in neural information processing systems, 29, 2016.
- Sequential neural likelihood: Fast likelihood-free inference with autoregressive flows. In The 22nd International Conference on Artificial Intelligence and Statistics, pp. 837–848. PMLR, 2019.
- Scalable diffusion models with transformers. arXiv preprint arXiv:2212.09748, 2022.
- Minimal hodgkin–huxley type models for different classes of cortical and thalamic neurons. Biological cybernetics, 99:427–441, 2008.
- Bayesflow: Learning complex stochastic models with invertible neural networks. IEEE transactions on neural networks and learning systems, 33(4):1452–1466, 2020.
- Jana: Jointly amortized neural approximation of complex Bayesian models. In Evans, R. J. and Shpitser, I. (eds.), Proceedings of the Thirty-Ninth Conference on Uncertainty in Artificial Intelligence, volume 216 of Proceedings of Machine Learning Research, pp. 1695–1706. PMLR, 31 Jul–04 Aug 2023.
- GATSBI: Generative adversarial training for simulation-based inference. In International Conference on Learning Representations, 2022.
- Arbitrary marginal neural ratio estimation for simulation-based inference. arXiv preprint arXiv:2110.00449, 2021.
- Score-based data assimilation. In Thirty-seventh Conference on Neural Information Processing Systems, 2023.
- A probabilistic state space model for joint inference from differential equations and data. In Ranzato, M., Beygelzimer, A., Dauphin, Y., Liang, P., and Vaughan, J. W. (eds.), Advances in Neural Information Processing Systems, volume 34, pp. 12374–12385. Curran Associates, Inc., 2021.
- Fuse it or lose it: Deep fusion for multimodal simulation-based inference, 2023.
- A survey on principles, models and methods for learning from irregularly sampled time series, 2021.
- Neural score estimation: Likelihood-free inference with conditional score based diffusion models. In Fifth Symposium on Advances in Approximate Bayesian Inference, 2023.
- Sequential monte carlo without likelihoods. Proceedings of the National Academy of Sciences, 104(6):1760–1765, 2007.
- Denoising diffusion implicit models. In International Conference on Learning Representations, 2021a.
- Pseudoinverse-guided diffusion models for inverse problems. In International Conference on Learning Representations, 2022.
- Generative modeling by estimating gradients of the data distribution. In Wallach, H., Larochelle, H., Beygelzimer, A., d'Alché-Buc, F., Fox, E., and Garnett, R. (eds.), Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc., 2019.
- Score-based generative modeling through stochastic differential equations. In International Conference on Learning Representations, 2021b.
- Arbitrary conditional distributions with energy. Advances in Neural Information Processing Systems, 34:752–763, 2021.
- Posterior matching for arbitrary conditioning. Advances in Neural Information Processing Systems, 35:18088–18099, 2022.
- sbi: A toolkit for simulation-based inference. Journal of Open Source Software, 5(52):2505, 2020. doi: 10.21105/joss.02505.
- Attention is all you need. Advances in neural information processing systems, 30, 2017.
- Volterra, V. Fluctuations in the abundance of a species considered mathematically. Nature, 118(2972):558–560, 1926.
- Missing data in amortized simulation-based neural posterior estimation. bioRxiv, 2023. doi: 10.1101/2023.01.09.523219.
- Faithful inversion of generative models for effective amortized inference. In Bengio, S., Wallach, H., Larochelle, H., Grauman, K., Cesa-Bianchi, N., and Garnett, R. (eds.), Advances in Neural Information Processing Systems, volume 31. Curran Associates, Inc., 2018.
- Structured conditional continuous normalizing flows for efficient amortized inference in graphical models. In International Conference on Artificial Intelligence and Statistics, pp. 4441–4451. PMLR, 2020.
- Graphically structured diffusion models. In Krause, A., Brunskill, E., Cho, K., Engelhardt, B., Sabato, S., and Scarlett, J. (eds.), Proceedings of the 40th International Conference on Machine Learning, volume 202 of Proceedings of Machine Learning Research, pp. 36887–36909. PMLR, 23–29 Jul 2023.
- Flow matching for scalable simulation-based inference. In Thirty-seventh Conference on Neural Information Processing Systems, 2023.
- Yadan, O. Hydra - a framework for elegantly configuring complex applications. Github, 2019.
- Zhang, J. Causal reasoning with ancestral graphs. Journal of Machine Learning Research, 9(7), 2008.
- Manuel Gloeckler (6 papers)
- Michael Deistler (16 papers)
- Christian Weilbach (10 papers)
- Frank Wood (98 papers)
- Jakob H. Macke (39 papers)