Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
184 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Diffusion posterior sampling for simulation-based inference in tall data settings (2404.07593v2)

Published 11 Apr 2024 in stat.ML, cs.LG, and stat.ME

Abstract: Determining which parameters of a non-linear model best describe a set of experimental data is a fundamental problem in science and it has gained much traction lately with the rise of complex large-scale simulators. The likelihood of such models is typically intractable, which is why classical MCMC methods can not be used. Simulation-based inference (SBI) stands out in this context by only requiring a dataset of simulations to train deep generative models capable of approximating the posterior distribution that relates input parameters to a given observation. In this work, we consider a tall data extension in which multiple observations are available to better infer the parameters of the model. The proposed method is built upon recent developments from the flourishing score-based diffusion literature and allows to estimate the tall data posterior distribution, while simply using information from a score network trained for a single context observation. We compare our method to recently proposed competing approaches on various numerical experiments and demonstrate its superiority in terms of numerical stability and computational cost.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (43)
  1. A stochastic version of the Jansen and Rit neural mass model: Analysis and numerics. The Journal of Mathematical Neuroscience, 7(1).
  2. Modeling brain dynamics in brain tumor patients using the virtual brain. eNeuro, 5(3). Society for Neuroscience.
  3. On markov chain monte carlo methods for tall data. Journal of Machine Learning Research, 18:1–43.
  4. Conditional image generation with score-based diffusion models christian etmann deep render.
  5. Bhatia, R. (2006). Positive definite matrices. Princeton Series in Applied Mathematics. Princeton University Press, Princeton, NJ.
  6. Pyro: Deep universal probabilistic programming. J. Mach. Learn. Res., 20:28:1–28:6.
  7. Tweedie moment projected diffusions for inverse problems. arXiv preprint arXiv:2310.06721.
  8. The tamed unadjusted langevin algorithm.
  9. Spectral density-based and measure-preserving ABC for partially observed diffusion processes. an illustration on hamiltonian SDEs. Statistics and Computing, 30(3):627–648.
  10. The frontier of simulation-based inference. Proceedings of the National Academy of Sciences (PNAS), 117:30055–30062.
  11. Neural importance sampling for rapid and reliable gravitational-wave inference. Phys. Rev. Lett., 130:171403.
  12. Compositional score modeling for simulation-based inference.
  13. Training deep neural density estimators to identify mechanistic models of neural dynamics. eLife, 9:e56261.
  14. Generative adversarial nets. In Ghahramani, Z., Welling, M., Cortes, C., Lawrence, N., and Weinberger, K., editors, Advances in Neural Information Processing Systems, volume 27. Curran Associates, Inc.
  15. Automatic posterior transformation for likelihood-free inference. In Chaudhuri, K. and Salakhutdinov, R., editors, Proceedings of the 36th International Conference on Machine Learning, volume 97, pages 2404–2414. PMLR.
  16. Array programming with NumPy. 585(7825):357–362. Number: 7825 Publisher: Nature Publishing Group.
  17. Likelihood-free MCMC with amortized approximate ratio estimators. In III, H. D. and Singh, A., editors, Proceedings of the 37th International Conference on Machine Learning, volume 119 of Proceedings of Machine Learning Research, pages 4239–4248. PMLR.
  18. Denoising diffusion probabilistic models. Advances in neural information processing systems, 33:6840–6851.
  19. Classifier-free diffusion guidance.
  20. Hunter, J. D. (2007). Matplotlib: A 2d graphics environment. 9(3):90–95. Conference Name: Computing in Science & Engineering.
  21. Estimation of non-normalized statistical models by score matching. Journal of Machine Learning Research, 6(4).
  22. Electroencephalogram and visual evoked potential generation in a mathematical model of coupled cortical columns. Biological Cybernetics 1995 73:4, 73:357–366.
  23. Elucidating the design space of diffusion-based generative models. In Proc. NeurIPS.
  24. Benchmarking simulation-based inference. In Banerjee, A. and Fukumizu, K., editors, Proceedings of The 24th International Conference on Artificial Intelligence and Statistics, volume 130 of Proceedings of Machine Learning Research, pages 343–351. PMLR.
  25. Normalizing flows for probabilistic modeling and inference. Journal of Machine Learning Research, 22:1–64.
  26. Sequential neural likelihood: Fast likelihood-free inference with autoregressive flows. 89:837–848.
  27. PyTorch: An imperative style, high-performance deep learning library. In Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc.
  28. Composable effects for flexible and accelerated probabilistic programming in numpyro. arXiv preprint arXiv:1912.11554.
  29. Exponential convergence of langevin distributions and their discrete approximations. Bernoulli, pages 341–363.
  30. HNPE: Leveraging Global Parameters for Neural Posterior Estimation. In NeurIPS 2021, Sydney (Online), Australia.
  31. The virtual brain: a simulator of primate brain network dynamics. Frontiers in Neuroinformatics, 7:10.
  32. Sequential neural score estimation: Likelihood-free inference with conditional score based diffusion models.
  33. Denoising diffusion implicit models. In International Conference on Learning Representations.
  34. Pseudoinverse-guided diffusion models for inverse problems. In International Conference on Learning Representations.
  35. Generative modeling by estimating gradients of the data distribution. Advances in neural information processing systems, 32.
  36. Score-based generative modeling through stochastic differential equations. In International Conference on Learning Representations.
  37. Tarantola, A. (2005). Inverse Problem Theory and Methods for Model Parameter Estimation. Society for Industrial and Applied Mathematics.
  38. Neural posterior estimation for exoplanetary atmospheric retrieval. Astronomy and Astrophysics, 672:A147.
  39. Vincent, P. (2011). A connection between score matching and denoising autoencoders. Neural computation, 23(7):1661–1674.
  40. SciPy 1.0: fundamental algorithms for scientific computing in python. 17(3):261–272. Number: 3 Publisher: Nature Publishing Group.
  41. Waskom, M. L. (2021). seaborn: statistical data visualization. 6(60):3021.
  42. Diffusion models: A comprehensive survey of methods and applications. ACM Computing Surveys, 56(4):1–39.
  43. Deep sets. In Guyon, I., Luxburg, U. V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., and Garnett, R., editors, Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc.
Citations (2)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com