Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
133 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Operator learning with PCA-Net: upper and lower complexity bounds (2303.16317v5)

Published 28 Mar 2023 in cs.LG, cs.NA, math.NA, and stat.ML

Abstract: PCA-Net is a recently proposed neural operator architecture which combines principal component analysis (PCA) with neural networks to approximate operators between infinite-dimensional function spaces. The present work develops approximation theory for this approach, improving and significantly extending previous work in this direction: First, a novel universal approximation result is derived, under minimal assumptions on the underlying operator and the data-generating distribution. Then, two potential obstacles to efficient operator learning with PCA-Net are identified, and made precise through lower complexity bounds; the first relates to the complexity of the output distribution, measured by a slow decay of the PCA eigenvalues. The other obstacle relates to the inherent complexity of the space of operators between infinite-dimensional input and output spaces, resulting in a rigorous and quantifiable statement of a "curse of parametric complexity", an infinite-dimensional analogue of the well-known curse of dimensionality encountered in high-dimensional approximation problems. In addition to these lower bounds, upper complexity bounds are finally derived. A suitable smoothness criterion is shown to ensure an algebraic decay of the PCA eigenvalues. Furthermore, it is shown that PCA-Net can overcome the general curse for specific operators of interest, arising from the Darcy flow and the Navier-Stokes equations.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (57)
  1. A general approximation lower bound in lpsuperscript𝑙𝑝l^{p}italic_l start_POSTSUPERSCRIPT italic_p end_POSTSUPERSCRIPT norm, with applications to feed-forward neural networks. Advances in Neural Information Processing Systems, 35:22396–22408, 2022.
  2. Neural Operator: Graph Kernel Network for Partial Differential Equations. In ICLR 2020 Workshop on Integration of Deep Neural Models and Differential Equations, 2020. URL https://openreview.net/forum?id=fg2ZFmXFO3.
  3. Model Reduction And Neural Networks For Parametric PDEs. The SMAI journal of computational mathematics, 7:121–157, 2021. doi: 10.5802/smai-jcm.74. URL https://smai-jcm.centre-mersenne.org/articles/10.5802/smai-jcm.74/.
  4. Learning elliptic partial differential equations with randomized linear algebra. Foundations of Computational Mathematics, pages 1–31, 2022.
  5. DeepM&Mnet: Inferring the electroconvection multiphysics fields based on operator approximation by neural networks. Journal of Computational Physics, 436:110296, 2021. Publisher: Elsevier.
  6. Optimal rates for the regularized least-squares algorithm. Foundations of Computational Mathematics, 7(3):331–368, 2007.
  7. Universal approximation to nonlinear operators by neural networks with arbitrary activation functions and its application to dynamical systems. IEEE Transactions on Neural Networks, 6(4):911–917, 1995.
  8. On the representation of solutions to elliptic pdes in barron spaces. Advances in neural information processing systems, 34:6454–6465, 2021.
  9. Algorithm for overcoming the curse of dimensionality for time-dependent non-convex Hamilton–Jacobi equations arising from optimal control and differential games problems. Journal of Scientific Computing, 73(2):617–643, 2017.
  10. Algorithm for overcoming the curse of dimensionality for state-dependent Hamilton-Jacobi equations. Journal of Computational Physics, 387:376–409, 2019.
  11. Convergence rates of best n-term galerkin approximations for a class of elliptic sPDEs. Foundations of Computational Mathematics, 10(6):615–646, 2010.
  12. Analytic regularity and polynomial approximation of parametric and stochastic elliptic PDE’s. Analysis and Applications, 9(01):11–47, 2011.
  13. Algorithms for overcoming the curse of dimensionality for certain Hamilton–Jacobi equations arising in control theory and elsewhere. Research in the Mathematical Sciences, 3(1):1–26, 2016.
  14. Overcoming the curse of dimensionality for some Hamilton–Jacobi partial differential equations via neural network architectures. Research in the Mathematical Sciences, 7(3):1–50, 2020.
  15. The cost-accuracy trade-off in operator learning with neural networks. To appear, Journal of Machine Learning, arXiv:2203.13181, 2022.
  16. Convergence rates for learning linear operators from noisy data. SIAM/ASA Journal on Uncertainty Quantification, 11(2):480–513, 2023.
  17. DeepONet prediction of linear instability waves in high-speed boundary layers. arXiv preprint arXiv:2105.08697, 2021. URL https://arxiv.org/abs/2105.08697.
  18. Lawrence C Evans. Partial differential equations, volume 19. American Mathematical Society, Providence, Rhode Island, 2nd edition, 2010.
  19. Deep Learning. MIT press, 2016.
  20. Multiwavelet-based operator learning for differential equations. Advances in Neural Information Processing Systems, 34:24048–24062, 2021.
  21. Non-intrusive reduced order modeling of nonlinear problems using neural networks. Journal of Computational Physics, 363:55–78, 2018. Publisher: Elsevier.
  22. Mionet: learning multiple-input operators via tensor product. SIAM Journal on Scientific Computing, 44(6):A3490–A3514, 2022. doi: 10.1137/22M1477751. URL https://doi.org/10.1137/22M1477751.
  23. Ian T Jolliffe. Principal component analysis. Springer Series in Statistics. Springer New York, NY, 2002.
  24. On universal approximation and error bounds for fourier neural operators. Journal of Machine Learning Research, 22(290):1–76, 2021. URL http://jmlr.org/papers/v22/21-0806.html.
  25. Neural operator: Learning maps between function spaces with applications to pdes. J. Mach. Learn. Res., 24(89):1–97, 2023.
  26. A theoretical analysis of deep neural networks and parametric PDEs. Constructive Approximation, 55(1):73–125, Feb 2022. ISSN 1432-0940. doi: 10.1007/s00365-021-09551-4. URL https://doi.org/10.1007/s00365-021-09551-4.
  27. The curse of dimensionality in operator learning. arXiv preprint arXiv:2306.15924, 2023.
  28. Error estimates for DeepONets: a deep learning framework in infinite dimensions. Transactions of Mathematics and Its Applications, 6(1), 03 2022. ISSN 2398-4945. doi: 10.1093/imatrm/tnac001. URL https://doi.org/10.1093/imatrm/tnac001. tnac001.
  29. The nonlocal neural operator: Universal approximation. (in preparation), 2023a.
  30. Nonlinear reconstruction for operator learning of PDEs with discontinuities. In 11th International Conference on Learning Representations (ICLR). OpenReview.net, 2023b.
  31. Multipole graph neural operator for parametric partial differential equations. In H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems (NeurIPS), volume 33, pages 6755–6766. Curran Associates, Inc., 2020.
  32. Fourier neural operator for parametric partial differential equations. In International Conference on Learning Representations, 2021. URL https://openreview.net/forum?id=c8P9NQVtmnO.
  33. Learning nonlinear operators via deeponet based on the universal approximation theorem of operators. Nature Machine Intelligence, 3(3):218–229, 2021.
  34. Vorticity and incompressible flow. Vorticity and Incompressible Flow, page 558, 2001.
  35. DeepMandMnet for hypersonics: Predicting the coupled flow and finite-rate chemistry behind a normal shock using neural-network approximation of operators. Journal of Computational Physics, 447:110698, 2021. ISSN 0021-9991. doi: https://doi.org/10.1016/j.jcp.2021.110698. URL https://www.sciencedirect.com/science/article/pii/S0021999121005933.
  36. Parametric complexity bounds for approximating pdes with neural networks. Advances in Neural Information Processing Systems, 34:15044–15055, 2021.
  37. Neural network approximations of pdes beyond linearity: A representational perspective. In International Conference on Machine Learning, pages 24139–24172. PMLR, 2023.
  38. Concentration inequalities under sub-Gaussian and sub-exponential conditions. Advances in Neural Information Processing Systems, 34:7588–7597, 2021.
  39. High-probability bounds for the reconstruction error of PCA. Statistics & Probability Letters, 161:108741, 2020.
  40. The random feature model for input-output maps between banach spaces. SIAM Journal on Scientific Computing, 43(5):A3212–A3243, 2021.
  41. Deep learning in high dimension: ReLU network expression rates for bayesian PDE inversion. Technical Report 2020-47, Seminar for Applied Mathematics, ETH ZĂĽrich, Switzerland, 2020.
  42. Exponential relu dnn expression of holomorphic maps in high dimension. Constructive Approximation, 55(1):537–582, 2022.
  43. Variationally mimetic operator networks. arXiv preprint arXiv:2209.12871, 2022.
  44. Remarks on inequalities for large deviation probabilities. Theory of Probability & Its Applications, 30(1):143–148, 1986.
  45. Variable-input deep operator networks. arXiv preprint arXiv:2205.11404, 2022.
  46. Convolutional neural operators. In ICLR 2023 Workshop on Physics for Machine Learning, 2023.
  47. Nonasymptotic upper bounds for the reconstruction error of PCA. The Annals of Statistics, 48(2):1098 – 1123, 2020. doi: 10.1214/19-AOS1839. URL https://doi.org/10.1214/19-AOS1839.
  48. Generalization properties of learning with random features. Advances in neural information processing systems, 30, 2017.
  49. Error analysis for deep neural network approximations of parametric hyperbolic conservation laws. arXiv preprint arXiv:2207.07362, 2021.
  50. Generic bounds on the approximation error for physics-informed (and) operator learning. In Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho, editors, Advances in Neural Information Processing Systems, 2022. URL https://openreview.net/forum?id=bF4eYy3LTR9.
  51. Deep learning in high dimension: Neural network expression rates for generalized polynomial chaos expansions in uq. Analysis and Applications, 17(01):19–55, 2019.
  52. NOMAD: Nonlinear manifold decoders for operator learning. In Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho, editors, Advances in Neural Information Processing Systems, 2022. URL https://openreview.net/forum?id=5OWV-sZvMl.
  53. Jonathan W Siegel. Optimal approximation rates for deep relu neural networks on sobolev spaces. arXiv preprint arXiv:2211.14400, 2022.
  54. Wavelet neural operator for solving parametric partial differential equations in computational mechanics problems. Computer Methods in Applied Mechanics and Engineering, 404:115783, 2023. ISSN 0045-7825. doi: https://doi.org/10.1016/j.cma.2022.115783. URL https://www.sciencedirect.com/science/article/pii/S0045782522007393.
  55. Roman Vershynin. High-dimensional probability: An introduction with applications in data science, volume 47. Cambridge university press, 2018.
  56. Dmitry Yarotsky. Error bounds for approximations with deep ReLU networks. Neural Networks, 94:103–114, 2017. Publisher: Elsevier.
  57. Dmitry Yarotsky. Optimal approximation of continuous functions by very deep relu networks. In Conference on learning theory, pages 639–649. PMLR, 2018.
Citations (22)

Summary

We haven't generated a summary for this paper yet.