Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
126 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A generative flow for conditional sampling via optimal transport (2307.04102v1)

Published 9 Jul 2023 in stat.ML, cs.LG, math.OC, and stat.CO

Abstract: Sampling conditional distributions is a fundamental task for Bayesian inference and density estimation. Generative models, such as normalizing flows and generative adversarial networks, characterize conditional distributions by learning a transport map that pushes forward a simple reference (e.g., a standard Gaussian) to a target distribution. While these approaches successfully describe many non-Gaussian problems, their performance is often limited by parametric bias and the reliability of gradient-based (adversarial) optimizers to learn these transformations. This work proposes a non-parametric generative model that iteratively maps reference samples to the target. The model uses block-triangular transport maps, whose components are shown to characterize conditionals of the target distribution. These maps arise from solving an optimal transport problem with a weighted $L2$ cost function, thereby extending the data-driven approach in [Trigila and Tabak, 2016] for conditional sampling. The proposed approach is demonstrated on a two dimensional example and on a parameter inference problem involving nonlinear ODEs.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (53)
  1. J. Adler and O. Öktem. Deep bayesian inversion. arXiv preprint arXiv:1811.05910, 2018.
  2. Input convex neural networks. In International Conference on Machine Learning, pages 146–155. PMLR, 2017.
  3. Wasserstein generative adversarial networks. In International conference on machine learning, pages 214–223. PMLR, 2017.
  4. On the representation and learning of monotone triangular transport maps. arXiv preprint arXiv:2009.10303, 2020.
  5. Conditional sampling with monotone GANs: from generative models to likelihood-free inference. arXiv preprint arXiv:2006.06755, 2023.
  6. The rate of convergence for approximate Bayesian computation. Electronic Journal of Statistics, 9(1):80 – 105, 2015.
  7. Conditional image generation with score-based diffusion models. arXiv preprint arXiv:2111.13606, 2021.
  8. Variational inference: A review for statisticians. Journal of the American statistical Association, 112(518):859–877, 2017.
  9. Y. Brenier. Polar factorization and monotone rearrangement of vector-valued functions. Communications on pure and applied mathematics, 44(4):375–417, 1991.
  10. Greedy inference with structure-exploiting lazy maps. Advances in Neural Information Processing Systems, 33:8330–8342, 2020.
  11. Supervised training of conditional monge maps. Advances in Neural Information Processing Systems, 35:6859–6872, 2022.
  12. From knothe’s transport to brenier’s map and a continuation method for optimal transport. SIAM Journal on Mathematical Analysis, 41(6):2554–2576, 2010.
  13. Vector quantile regression: An optimal transport approach. Annals of Statistics, 44(3):1165–1192, 2016.
  14. A gradient descent solution to the Monge-Kantorovich problem. Applied Mathematical Sciences, 3(22):1071–1080, 2009.
  15. Svgd as a kernelized Wasserstein gradient flow of the chi-squared divergence. Advances in Neural Information Processing Systems, 33:2098–2109, 2020.
  16. The frontier of simulation-based inference. Proceedings of the National Academy of Sciences, 117(48):30055–30062, 2020.
  17. M. Cuturi. Sinkhorn distances: Lightspeed computation of optimal transport. Advances in neural information processing systems, 26, 2013.
  18. A. Figalli and F. Glaudo. An Invitation to Optimal Transport, Wasserstein Distances, and Gradient Flows. EMS Press, 2021. 10.4171/ETB/22.
  19. W. Gangbo. An elementary proof of the polar factorization of vector-valued functions. Archive for rational mechanics and analysis, 128:381–399, 1994.
  20. FFJORD: Free-form continuous dynamics for scalable reversible generative models. arXiv preprint arXiv:1810.01367, 2018.
  21. Automatic posterior transformation for likelihood-free inference. In International Conference on Machine Learning, pages 2404–2414. PMLR, 2019.
  22. Lipschitz regularized gradient flows and latent generative particles. arXiv preprint arXiv:2210.17230, 2022.
  23. H. Knothe. Contributions to the theory of convex bodies. Michigan Mathematical Journal, 4(1):39–52, 1957.
  24. Normalizing flows: An introduction and review of current methods. IEEE transactions on pattern analysis and machine intelligence, 43(11):3964–3979, 2020.
  25. Wasserstein generative learning of conditional distribution. arXiv preprint arXiv:2112.10039, 2021.
  26. Likelihood-free inference with emulator networks. In Symposium on Advances in Approximate Bayesian Inference, pages 32–53. PMLR, 2019.
  27. Optimal transport mapping via input convex neural networks. In International Conference on Machine Learning, pages 6672–6681. PMLR, 2020.
  28. Sampling via measure transport: An introduction. In Handbook of Uncertainty Quantification, pages 1–41. Springer International Publishing, Cham, 2016. ISBN 978-3-319-11259-6. 10.1007/978-3-319-11259-6_23-1.
  29. M. Mirza and S. Osindero. Conditional generative adversarial nets. arXiv preprint arXiv:1411.1784, 2014.
  30. E. A. Nadaraya. On estimating regression. Theory of Probability & Its Applications, 9(1):141–142, 1964.
  31. High-dimensional ABC. In Handbook of Approximate Bayesian Computation, pages 211–241. Chapman and Hall/CRC, 2018.
  32. OT-flow: Fast and accurate continuous normalizing flows via optimal transport. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35(10), pages 9223–9232, 2021.
  33. Normalizing flows for probabilistic modeling and inference. The Journal of Machine Learning Research, 22(1):2617–2680, 2021.
  34. Computational optimal transport: With applications to data science. Foundations and Trends® in Machine Learning, 11(5-6):355–607, 2019.
  35. A.-A. Pooladian and J. Niles-Weed. Entropic estimation of optimal transport maps. arXiv preprint arXiv:2109.12004, 2021.
  36. D. Rezende and S. Mohamed. Variational inference with normalizing flows. In International conference on machine learning, pages 1530–1538. PMLR, 2015.
  37. Monte Carlo statistical methods, volume 2. Springer, 1999.
  38. M. Rosenblatt. Remarks on a multivariate transformation. The annals of mathematical statistics, 23(3):470–472, 1952.
  39. L. Ruthotto and E. Haber. An introduction to deep generative modeling. GAMM-Mitteilungen, 44(2):e202100008, 2021.
  40. Image super-resolution via iterative refinement. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2022.
  41. F. Santambrogio. Optimal transport for applied mathematicians. Birkäuser, NY, 55(58-63):94, 2015.
  42. Handbook of approximate Bayesian computation. CRC Press, 2018.
  43. Coupling techniques for nonlinear ensemble filtering. SIAM Review, 64(4):921–953, 2022.
  44. A family of nonparametric density estimation algorithms. Communications on Pure and Applied Mathematics, 66(2):145–164, 2013.
  45. E. G. Tabak and E. Vanden-Eijnden. Density estimation by dual ascent of the log-likelihood. Communications in Mathematical Sciences, 8(1):217–233, 2010.
  46. A. Taghvaei and B. Hosseini. An optimal transport formulation of Bayes’ law for nonlinear filtering algorithms. In 2022 IEEE 61st Conference on Decision and Control (CDC), pages 6608–6613. IEEE, 2022.
  47. G. Trigila and E. G. Tabak. Data-driven optimal transport. Communications on Pure and Applied Mathematics, 69(4):613–648, 2016.
  48. Conditional density estimation with Bayesian normalising flows. In Second workshop on Bayesian Deep Learning, 2017.
  49. T. Uscidda and M. Cuturi. The monge gap: A regularizer to learn all transport maps. arXiv preprint arXiv:2302.04953, 2023.
  50. C. Villani et al. Optimal transport: old and new, volume 338. Springer, 2009.
  51. G. S. Watson. Smooth regression analysis. Sankhyā: The Indian Journal of Statistics, Series A, pages 359–372, 1964.
  52. A. Wehenkel and G. Louppe. Unconstrained monotonic neural networks. Advances in neural information processing systems, 32, 2019.
  53. Learning likelihoods with conditional normalizing flows. arXiv preprint arXiv:1912.00042, 2019.
Citations (5)

Summary

We haven't generated a summary for this paper yet.