A generative flow for conditional sampling via optimal transport (2307.04102v1)
Abstract: Sampling conditional distributions is a fundamental task for Bayesian inference and density estimation. Generative models, such as normalizing flows and generative adversarial networks, characterize conditional distributions by learning a transport map that pushes forward a simple reference (e.g., a standard Gaussian) to a target distribution. While these approaches successfully describe many non-Gaussian problems, their performance is often limited by parametric bias and the reliability of gradient-based (adversarial) optimizers to learn these transformations. This work proposes a non-parametric generative model that iteratively maps reference samples to the target. The model uses block-triangular transport maps, whose components are shown to characterize conditionals of the target distribution. These maps arise from solving an optimal transport problem with a weighted $L2$ cost function, thereby extending the data-driven approach in [Trigila and Tabak, 2016] for conditional sampling. The proposed approach is demonstrated on a two dimensional example and on a parameter inference problem involving nonlinear ODEs.
- J. Adler and O. Öktem. Deep bayesian inversion. arXiv preprint arXiv:1811.05910, 2018.
- Input convex neural networks. In International Conference on Machine Learning, pages 146–155. PMLR, 2017.
- Wasserstein generative adversarial networks. In International conference on machine learning, pages 214–223. PMLR, 2017.
- On the representation and learning of monotone triangular transport maps. arXiv preprint arXiv:2009.10303, 2020.
- Conditional sampling with monotone GANs: from generative models to likelihood-free inference. arXiv preprint arXiv:2006.06755, 2023.
- The rate of convergence for approximate Bayesian computation. Electronic Journal of Statistics, 9(1):80 – 105, 2015.
- Conditional image generation with score-based diffusion models. arXiv preprint arXiv:2111.13606, 2021.
- Variational inference: A review for statisticians. Journal of the American statistical Association, 112(518):859–877, 2017.
- Y. Brenier. Polar factorization and monotone rearrangement of vector-valued functions. Communications on pure and applied mathematics, 44(4):375–417, 1991.
- Greedy inference with structure-exploiting lazy maps. Advances in Neural Information Processing Systems, 33:8330–8342, 2020.
- Supervised training of conditional monge maps. Advances in Neural Information Processing Systems, 35:6859–6872, 2022.
- From knothe’s transport to brenier’s map and a continuation method for optimal transport. SIAM Journal on Mathematical Analysis, 41(6):2554–2576, 2010.
- Vector quantile regression: An optimal transport approach. Annals of Statistics, 44(3):1165–1192, 2016.
- A gradient descent solution to the Monge-Kantorovich problem. Applied Mathematical Sciences, 3(22):1071–1080, 2009.
- Svgd as a kernelized Wasserstein gradient flow of the chi-squared divergence. Advances in Neural Information Processing Systems, 33:2098–2109, 2020.
- The frontier of simulation-based inference. Proceedings of the National Academy of Sciences, 117(48):30055–30062, 2020.
- M. Cuturi. Sinkhorn distances: Lightspeed computation of optimal transport. Advances in neural information processing systems, 26, 2013.
- A. Figalli and F. Glaudo. An Invitation to Optimal Transport, Wasserstein Distances, and Gradient Flows. EMS Press, 2021. 10.4171/ETB/22.
- W. Gangbo. An elementary proof of the polar factorization of vector-valued functions. Archive for rational mechanics and analysis, 128:381–399, 1994.
- FFJORD: Free-form continuous dynamics for scalable reversible generative models. arXiv preprint arXiv:1810.01367, 2018.
- Automatic posterior transformation for likelihood-free inference. In International Conference on Machine Learning, pages 2404–2414. PMLR, 2019.
- Lipschitz regularized gradient flows and latent generative particles. arXiv preprint arXiv:2210.17230, 2022.
- H. Knothe. Contributions to the theory of convex bodies. Michigan Mathematical Journal, 4(1):39–52, 1957.
- Normalizing flows: An introduction and review of current methods. IEEE transactions on pattern analysis and machine intelligence, 43(11):3964–3979, 2020.
- Wasserstein generative learning of conditional distribution. arXiv preprint arXiv:2112.10039, 2021.
- Likelihood-free inference with emulator networks. In Symposium on Advances in Approximate Bayesian Inference, pages 32–53. PMLR, 2019.
- Optimal transport mapping via input convex neural networks. In International Conference on Machine Learning, pages 6672–6681. PMLR, 2020.
- Sampling via measure transport: An introduction. In Handbook of Uncertainty Quantification, pages 1–41. Springer International Publishing, Cham, 2016. ISBN 978-3-319-11259-6. 10.1007/978-3-319-11259-6_23-1.
- M. Mirza and S. Osindero. Conditional generative adversarial nets. arXiv preprint arXiv:1411.1784, 2014.
- E. A. Nadaraya. On estimating regression. Theory of Probability & Its Applications, 9(1):141–142, 1964.
- High-dimensional ABC. In Handbook of Approximate Bayesian Computation, pages 211–241. Chapman and Hall/CRC, 2018.
- OT-flow: Fast and accurate continuous normalizing flows via optimal transport. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35(10), pages 9223–9232, 2021.
- Normalizing flows for probabilistic modeling and inference. The Journal of Machine Learning Research, 22(1):2617–2680, 2021.
- Computational optimal transport: With applications to data science. Foundations and Trends® in Machine Learning, 11(5-6):355–607, 2019.
- A.-A. Pooladian and J. Niles-Weed. Entropic estimation of optimal transport maps. arXiv preprint arXiv:2109.12004, 2021.
- D. Rezende and S. Mohamed. Variational inference with normalizing flows. In International conference on machine learning, pages 1530–1538. PMLR, 2015.
- Monte Carlo statistical methods, volume 2. Springer, 1999.
- M. Rosenblatt. Remarks on a multivariate transformation. The annals of mathematical statistics, 23(3):470–472, 1952.
- L. Ruthotto and E. Haber. An introduction to deep generative modeling. GAMM-Mitteilungen, 44(2):e202100008, 2021.
- Image super-resolution via iterative refinement. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2022.
- F. Santambrogio. Optimal transport for applied mathematicians. Birkäuser, NY, 55(58-63):94, 2015.
- Handbook of approximate Bayesian computation. CRC Press, 2018.
- Coupling techniques for nonlinear ensemble filtering. SIAM Review, 64(4):921–953, 2022.
- A family of nonparametric density estimation algorithms. Communications on Pure and Applied Mathematics, 66(2):145–164, 2013.
- E. G. Tabak and E. Vanden-Eijnden. Density estimation by dual ascent of the log-likelihood. Communications in Mathematical Sciences, 8(1):217–233, 2010.
- A. Taghvaei and B. Hosseini. An optimal transport formulation of Bayes’ law for nonlinear filtering algorithms. In 2022 IEEE 61st Conference on Decision and Control (CDC), pages 6608–6613. IEEE, 2022.
- G. Trigila and E. G. Tabak. Data-driven optimal transport. Communications on Pure and Applied Mathematics, 69(4):613–648, 2016.
- Conditional density estimation with Bayesian normalising flows. In Second workshop on Bayesian Deep Learning, 2017.
- T. Uscidda and M. Cuturi. The monge gap: A regularizer to learn all transport maps. arXiv preprint arXiv:2302.04953, 2023.
- C. Villani et al. Optimal transport: old and new, volume 338. Springer, 2009.
- G. S. Watson. Smooth regression analysis. Sankhyā: The Indian Journal of Statistics, Series A, pages 359–372, 1964.
- A. Wehenkel and G. Louppe. Unconstrained monotonic neural networks. Advances in neural information processing systems, 32, 2019.
- Learning likelihoods with conditional normalizing flows. arXiv preprint arXiv:1912.00042, 2019.