Adaptive operator learning for infinite-dimensional Bayesian inverse problems (2310.17844v3)
Abstract: The fundamental computational issues in Bayesian inverse problems (BIP) governed by partial differential equations (PDEs) stem from the requirement of repeated forward model evaluations. A popular strategy to reduce such costs is to replace expensive model simulations with computationally efficient approximations using operator learning, motivated by recent progress in deep learning. However, using the approximated model directly may introduce a modeling error, exacerbating the already ill-posedness of inverse problems. Thus, balancing between accuracy and efficiency is essential for the effective implementation of such approaches. To this end, we develop an adaptive operator learning framework that can reduce modeling error gradually by forcing the surrogate to be accurate in local areas. This is accomplished by adaptively fine-tuning the pre-trained approximate model with training points chosen by a greedy algorithm during the posterior evaluation process. To validate our approach, we use DeepOnet to construct the surrogate and unscented Kalman inversion (UKI) to approximate the BIP solution, respectively. Furthermore, we present a rigorous convergence guarantee in the linear case using the UKI framework. The approach is tested on a number of benchmarks, including the Darcy flow, the heat source inversion problem, and the reaction-diffusion problem. The numerical results show that our method can significantly reduce computational costs while maintaining inversion accuracy.
- Scalable posterior approximations for large-scale bayesian inverse problems via likelihood-informed parameter and state reduction. Journal of Computational Physics, 315:363–387, 2016.
- A bayesian approach to estimate uncertainty for full-waveform inversion using a priori information from depth migration. Geophysics, 81(5):R307–R323, 2016.
- A fast and scalable method for a-optimal design of experiments for infinite-dimensional bayesian nonlinear inverse problems. SIAM Journal on Scientific Computing, 38(1):A243–A272, 2016.
- A computational framework for infinite-dimensional bayesian inverse problems part i: The linearized case, with application to global seismic inversion. SIAM Journal on Scientific Computing, 35(6):A2494–A2523, 2013.
- A computational framework for infinite-dimensional bayesian inverse problems, part ii: Stochastic newton mcmc with application to ice sheet flow inverse problems. SIAM Journal on Scientific Computing, 36(4):A1525–A1555, 2014.
- Mcmc methods for functions: modifying old algorithms to make them faster. 2013.
- Ensemble samplers with affine invariance. Communications in applied mathematics and computational science, 5(1):65–80, 2010.
- Weak convergence and optimal scaling of random walk metropolis algorithms. The annals of applied probability, 7(1):110–120, 1997.
- Data-driven model reduction for the bayesian solution of inverse problems. International Journal for Numerical Methods in Engineering, 102(5):966–990, 2015.
- Parameter and state model reduction for large-scale statistical inverse problems. SIAM Journal on Scientific Computing, 32(5):2523–2542, 2010.
- Dimensionality reduction and polynomial chaos acceleration of bayesian inference in inverse problems. Journal of Computational Physics, 228(6):1862–1902, 2009.
- On the convergence of the laplace approximation and noise-level-robustness of laplace-based monte carlo methods for bayesian inverse problems. Numerische Mathematik, 145:915–971, 2020.
- Adaptive construction of surrogates for the bayesian solution of inverse problems. SIAM Journal on Scientific Computing, 36(3):A1163–A1186, 2014.
- Convergence analysis of surrogate-based methods for bayesian inverse problems. Inverse Problems, 33(12):125001, 2017.
- An adaptive surrogate modeling based on deep neural networks for large-scale bayesian inverse problems. Communications in Computational Physics, 28(5):2180–2205, 2020.
- Solving high-dimensional partial differential equations using deep learning. Proceedings of the National Academy of Sciences, 115(34):8505–8510, 2018.
- Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. Journal of Computational physics, 378:686–707, 2019.
- C. Schwab and J. Zech. Deep learning in high dimension: Neural network expression rates for generalized polynomial chaos expansions in uq. Analysis and Applications, 17(01):19–55, 2019.
- R. K. Tripathy and I. Bilionis. Deep UQ: Learning deep neural network surrogate models for high dimensional uncertainty quantification. Journal of Computational Physics, 375:565–588, 2018.
- Y. Zhu and N. Zabaras. Bayesian deep convolutional encoder–decoder networks for surrogate modeling and uncertainty quantification. Journal of Computational Physics, 366:415–447, 2018.
- A deep surrogate approach to efficient Bayesian inversion in PDE and integral equation models. arXiv:1910.01547, 2019.
- An acceleration strategy for randomize-then-optimize sampling via deep neural networks. Journal of Computational Mathematics, 39(6):848–864, 2021.
- Surrogate modeling for bayesian inverse problems based on physics-informed neural networks. Journal of Computational Physics, 475:111841, 2023.
- Adaptive Physics-Informed Neural Networks for Markov-Chain Monte Carlo. arXiv: 2008.01604, 2020.
- When and why pinns fail to train: A neural tangent kernel perspective. Journal of Computational Physics, 449:110768, 2022.
- Characterizing possible failure modes in physics-informed neural networks. Advances in Neural Information Processing Systems, 34:26548–26560, 2021.
- Failure-informed adaptive sampling for pinns. SIAM Journal on Scientific Computing, 45(4):A1971–A1994, 2023.
- Failure-informed adaptive sampling for pinns, part ii: combining with re-sampling and subset simulation. arXiv preprint arXiv:2302.01529, 2023.
- Self-adaptive physics-informed neural networks using a soft attention mechanism. arXiv preprint arXiv:2009.04544, 2020.
- Self-adaptive loss balanced physics-informed neural networks. Neurocomputing, 496:11–34, 2022.
- Fourier neural operator for parametric partial differential equations. arXiv preprint arXiv:2010.08895, 2020.
- Learning nonlinear operators via deeponet based on the universal approximation theorem of operators. Nature machine intelligence, 3(3):218–229, 2021.
- Residual-based error correction for neural operator accelerated infinite-dimensional bayesian inverse problems. Journal of Computational Physics, 486:112104, 2023.
- Solving inverse problems with deep neural networks–robustness included? IEEE transactions on pattern analysis and machine intelligence, 45(1):1119–1134, 2022.
- Iterated kalman methodology for inverse problems. Journal of Computational Physics, 463:111262, 2022.
- Andrew M Stuart. Inverse problems: a bayesian perspective. Acta numerica, 19:451–559, 2010.
- Handbook of Markov chain Monte Carlo. Chapman & Hall/CRC Handbooks of Modern Statistical Methods. CRC Press, Boca Raton, FL, 2011.
- Variational inference: A review for statisticians. Journal of the American statistical Association, 112(518):859–877, 2017.
- Error estimates for deeponets: A deep learning framework in infinite dimensions. Transactions of Mathematics and Its Applications, 6(1):1–141, 2022.
- Adaptive multi-fidelity polynomial chaos approach to bayesian inference in inverse problems. Journal of Computational Physics, 381:110–128, 2019.
- Projected stein variational newton: A fast and scalable bayesian inference method in high dimensions. In Advances in Neural Information Processing Systems, pages 15130–15139, 2019.
- A stein variational newton method. In Advances in Neural Information Processing Systems, pages 9169–9179, 2018.
- Interacting langevin diffusions: Gradient structure and ensemble kalman sampler. SIAM Journal on Applied Dynamical Systems, 19(1):412–441, 2020.
- Q. Liu and D. Wang. Stein variational gradient descent: A general purpose bayesian inference algorithm. In Advances in neural information processing systems, pages 2378–2386, 2016.
- Stein variational gradient descent with local approximations. Computer Methods in Applied Mechanics and Engineering, 386:114087, 2021.
- Convergence acceleration of ensemble Kalman inversion in nonlinear settings. Mathematics of computation, 91:1247–1280, 2021.
- Consensus-based sampling. Studies in Applied Mathematics, 148(3):1069–1140, 2022.
- Analysis of the ensemble and polynomial chaos Kalman filters in Bayesian inverse problems. SIAM/ASA Journal on Uncertainty Quantification, 3(1):823–851, 2015.
- Efficient derivative-free bayesian inference for large-scale inverse problems. Inverse Problems, 38(12):125006, 2022.
- Ensemble Kalman methods for inverse problems. Inverse Problems, 29(4):045001, 2013.
- Adaptive ensemble kalman inversion with statistical linearization. Communications in Computational Physics, 33(5):1357–1380, 2023.
- Adaptive Tikhonov strategies for stochastic ensemble Kalman inversion. Inverse Problems, 38(4):045009, 2022.
- An adaptive multifidelity PC-based ensemble kalman inversion for inverse problems. International Journal for Uncertainty Quantification, 9(3):205–220, 2019.
- Eric A Wan and Rudolph Van Der Merwe. The unscented kalman filter for nonlinear estimation. In Proceedings of the IEEE 2000 Adaptive Systems for Signal Processing, Communications, and Control Symposium (Cat. No. 00EX373), pages 153–158. Ieee, 2000.
- Zhiwei Gao (11 papers)
- Liang Yan (18 papers)
- Tao Zhou (398 papers)