Feynman-Kac Operator Expectation Estimator
Abstract: The Feynman-Kac Operator Expectation Estimator (FKEE) is an innovative method for estimating the target Mathematical Expectation $\mathbb{E}_{X\sim P}[f(X)]$ without relying on a large number of samples, in contrast to the commonly used Markov Chain Monte Carlo (MCMC) Expectation Estimator. FKEE comprises diffusion bridge models and approximation of the Feynman-Kac operator. The key idea is to use the solution to the Feynmann-Kac equation at the initial time $u(x_0,0)=\mathbb{E}[f(X_T)|X_0=x_0]$. We use Physically Informed Neural Networks (PINN) to approximate the Feynman-Kac operator, which enables the incorporation of diffusion bridge models into the expectation estimator and significantly improves the efficiency of using data while substantially reducing the variance. Diffusion Bridge Model is a more general MCMC method. In order to incorporate extensive MCMC algorithms, we propose a new diffusion bridge model based on the Minimum Wasserstein distance. This diffusion bridge model is universal and reduces the training time of the PINN. FKEE also reduces the adverse impact of the curse of dimensionality and weakens the assumptions on the distribution of $X$ and performance function $f$ in the general MCMC expectation estimator. The theoretical properties of this universal diffusion bridge model are also shown. Finally, we demonstrate the advantages and potential applications of this method through various concrete experiments, including the challenging task of approximating the partition function in the random graph model such as the Ising model.
- Banach wasserstein gan. Advances In Neural Information Processing Systems, 31, 2018.
- S. Ejaz Ahmed. Markov Chain Monte Carlo: Stochastic Simulation for Bayesian Inference. Technometrics, 50:97 – 97, 2008.
- Building normalizing flows with stochastic interpolants. In The Eleventh International Conference on Learning Representations, 2022.
- Numerically solving parametric families of high-dimensional kolmogorov partial differential equations via deep learning. Advances in Neural Information Processing Systems, 33:16615–16627, 2020.
- On parameter estimation with the Wasserstein distance. Information and Inference: A Journal of the IMA, 2017.
- MCMC methods for diffusion bridges. Stochastics and Dynamics, 8(03):319–350, 2008.
- Corrigendum to “Simple simulation of diffusion bridges with application to likelihood inference for diffusions”. Bernoulli, 2010.
- Three ways to solve partial differential equations with neural networks — A review. GAMM-Mitteilungen, 44, 2021.
- Russel E. Caflisch. Monte Carlo and quasi-Monte Carlo methods. Acta Numerica, 7:1 – 49, 1998.
- Stochastic Gradient Hamiltonian Monte Carlo. In International Conference on Machine Learning, 2014.
- Adaptive trajectories sampling for solving pdes with deep learning methods. ArXiv, abs/2303.15704, 2023.
- Convergence of Langevin MCMC in KL-divergence. In Algorithmic Learning Theory, pages 186–211. PMLR, 2018.
- Underdamped Langevin MCMC: A non-asymptotic analysis. In Conference on learning theory, pages 300–323. PMLR, 2018.
- Barry A Cipra. An introduction to the Ising model. The American Mathematical Monthly, 94(10):937–959, 1987.
- Marco Cuturi. Sinkhorn distances: Lightspeed computation of optimal transport. Advances in Neural Information Processing Systems, 26, 2013.
- Error analysis for physics-informed neural networks (pinns) approximating kolmogorov pdes. Advances in Computational Mathematics, 48(6):79, 2022.
- Feynman-kac formulae. Springer, 2004.
- Non-convex Learning via Replica Exchange Stochastic Gradient MCMC. Proceedings of Machine Learning Research, 119:2474–2483, 2020.
- Diffusion models beat gans on image synthesis. Advances in Neural Information Processing Systems, 34:8780–8794, 2021.
- Learning effective stochastic differential equations from microscopic simulations: Linking stochastic numerics to deep learning. Chaos: An Interdisciplinary Journal of Nonlinear Science, 33(2), 2023.
- Perfect sampling: a review and applications to signal processing. IEEE Trans. Signal Process., 50:345–356, 2002.
- Random graph modeling: A survey of the concepts. ACM computing surveys (CSUR), 52(6):1–36, 2019.
- Till Daniel Frank. Nonlinear Fokker-Planck equations: fundamentals and applications. Springer Science & Business Media, 2005.
- Langevin diffusions on the torus: estimation and applications. Statistics and Computing, 29:1–22, 2017.
- MCMC Variational Inference via Uncorrected Hamiltonian Annealing. In Neural Information Processing Systems, 2021.
- Fast Doubly-Adaptive MCMC to Estimate the Gibbs Partition Function with Weak Mixing Time Bounds. Advances in Neural Information Processing Systems, 34:25760–25772, 2021.
- Tim Hesterberg. Monte Carlo Strategies in Scientific Computing. Technometrics, 44:403 – 404, 2002.
- Flow++: Improving flow-based generative models with variational dequantization and architecture design. In International Conference on Machine Learning, pages 2722–2730. PMLR, 2019.
- Unbiased Markov chain Monte Carlo with couplings. arXiv: Methodology, 2017.
- Efficient and accurate gradients for neural sdes. Advances in Neural Information Processing Systems, 34:18747–18761, 2021.
- Variational diffusion models. Advances in Neural Information Processing Systems, 34:21696–21707, 2021.
- Wuchen Li. Langevin dynamics for the probability of Markov jumping processes. arXiv preprint arXiv:2307.00678, 2023.
- On extension of the Markov chain approximation method for computing Feynman–Kac type expectations. arXiv preprint arXiv:2302.11698, 2023.
- Let us Build Bridges: Understanding and Extending Diffusion Generative Models. ArXiv, abs/2208.14699, 2022.
- Gábor Lugosi. Concentration Inequalities. In Annual Conference Computational Learning Theory, 2003.
- Deep energy-based modeling of discrete-time physics. Advances in Neural Information Processing Systems, 33:13100–13111, 2020.
- Stochastic Numerics for Mathematical Physics. Scientific Computation, 2004.
- High-order Langevin diffusion yields an accelerated MCMC algorithm. The Journal of Machine Learning Research, 22(1):1919–1959, 2021.
- Neural Importance Sampling. ACM Transactions on Graphics (TOG), 38:1 – 19, 2018.
- Random graph models of social networks. Proceedings of the national academy of sciences, 99(suppl_1):2566–2572, 2002.
- Improved denoising diffusion probabilistic models. In International Conference on Machine Learning, pages 8162–8171. PMLR, 2021.
- Control functionals for Monte Carlo integration. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 79, 2014.
- Shige Peng. Nonlinear Expectations and Stochastic Calculus under Uncertainty. arXiv e-prints, pages arXiv–1002, 2010.
- Huyên Pham. Feynman-Kac Representation of Fully Nonlinear PDEs and Applications. Acta Mathematica Vietnamica, 40:255–269, 2014.
- Eckhard Platen. An introduction to numerical methods for stochastic differential equations. Acta numerica, 8:197–246, 1999.
- Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. J. Comput. Phys., 378:686–707, 2019.
- Black box variational inference. In Artificial intelligence and statistics, pages 814–822. PMLR, 2014.
- Fokker-planck equation. Springer, 1996.
- R Tyrrell Rockafellar and Roger J-B Wets. Variational analysis, volume 317. Springer Science & Business Media, 2009.
- Filippo Santambrogio. Optimal transport for applied mathematicians. Birkäuser, NY, 55(58-63):94, 2015.
- Applied stochastic differential equations, volume 10. Cambridge University Press, 2019.
- Accelerated Training of Physics-Informed Neural Networks (PINNs) using Meshless Discretizations. Advances in Neural Information Processing Systems, 35:1034–1046, 2022.
- Score-Based Generative Modeling through Stochastic Differential Equations. In International Conference on Learning Representations, 2020.
- Semi-Exact Control Functionals From Sard’s Method. Biometrika, 2020.
- Gradient-free Hamiltonian Monte Carlo with Efficient Kernel Exponential Families. In Advances in Neural Information Processing Systems, 2015.
- Discrete flows: Invertible generative models of discrete data. Advances in Neural Information Processing Systems, 32, 2019.
- Neural Stochastic Differential Equations: Deep Latent Gaussian Models in the Diffusion Limit. ArXiv, abs/1905.09883, 2019.
- Cédric Villani. Optimal Transport: Old and New. Springer, 2008.
- Langevin diffusions and the Metropolis-adjusted Langevin algorithm. Statistics & Probability Letters, 91:14–19, 2013.
- Improved variational autoencoders for text modeling using dilated convolutions. In International Conference on Machine Learning, pages 3881–3890. PMLR, 2017.
- A-PINN: Auxiliary physics informed neural networks for forward and inverse problems of nonlinear integro-differential equations. J. Comput. Phys., 462:111260, 2022.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.