Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
125 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Adaptive Finite Element Interpolated Neural Networks (2403.14054v1)

Published 21 Mar 2024 in math.NA and cs.NA

Abstract: The use of neural networks to approximate partial differential equations (PDEs) has gained significant attention in recent years. However, the approximation of PDEs with localised phenomena, e.g., sharp gradients and singularities, remains a challenge, due to ill-defined cost functions in terms of pointwise residual sampling or poor numerical integration. In this work, we introduce $h$-adaptive finite element interpolated neural networks. The method relies on the interpolation of a neural network onto a finite element space that is gradually adapted to the solution during the training process to equidistribute a posteriori error indicator. The use of adaptive interpolation is essential in preserving the non-linear approximation capabilities of the neural networks to effectively tackle problems with localised features. The training relies on a gradient-based optimisation of a loss function based on the (dual) norm of the finite element residual of the interpolated neural network. Automatic mesh adaptation (i.e., refinement and coarsening) is performed based on a posteriori error indicators till a certain level of accuracy is reached. The proposed methodology can be applied to indefinite and nonsymmetric problems. We carry out a detailed numerical analysis of the scheme and prove several a priori error estimates, depending on the expressiveness of the neural network compared to the interpolation mesh. Our numerical experiments confirm the effectiveness of the method in capturing sharp gradients and singularities for forward PDE problems, both in 2D and 3D scenarios. We also show that the proposed preconditioning strategy (i.e., using a dual residual norm of the residual as a cost function) enhances training robustness and accelerates convergence.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (52)
  1. “Finite Elements I” Springer International Publishing, 2021 DOI: 10.1007/978-3-030-56341-7
  2. Douglas N. Arnold, Richard S. Falk and Ragnar Winther “Finite element exterior calculus, homological techniques, and applications” In Acta Numerica 15 Cambridge University Press (CUP), 2006, pp. 1–155 DOI: 10.1017/s0962492906210018
  3. “Composing Scalable Nonlinear Algebraic Solvers” In SIAM Review 57.4 Society for Industrial & Applied Mathematics (SIAM), 2015, pp. 535–565 DOI: 10.1137/130936725
  4. Santiago Badia, Alberto F. Martín and Javier Principe “Multilevel Balancing Domain Decomposition at Extreme Scales” In SIAM Journal on Scientific Computing 38.1 Society for Industrial & Applied Mathematics (SIAM), 2016, pp. C22–C52 DOI: 10.1137/15m1013511
  5. “Scheduling Massively Parallel Multigrid for Multilevel Monte Carlo Methods” In SIAM Journal on Scientific Computing 39.5 Society for Industrial & Applied Mathematics (SIAM), 2017, pp. S873–S897 DOI: 10.1137/16m1083591
  6. “A posteriori error estimation in finite element analysis” In Computer Methods in Applied Mechanics and Engineering 142.1, 1997, pp. 1–88 DOI: https://doi.org/10.1016/S0045-7825(96)01107-3
  7. Joachim Schöberl “A Posteriori Error Estimates for Maxwell Equations” In Mathematics of Computation 77.262 American Mathematical Society, 2008, pp. 633–649 URL: http://www.jstor.org/stable/40234527
  8. Olaf Ronneberger, Philipp Fischer and Thomas Brox “U-Net: Convolutional Networks for Biomedical Image Segmentation” In Medical Image Computing and Computer-Assisted Intervention – MICCAI 2015 Cham: Springer International Publishing, 2015, pp. 234–241 DOI: https://doi.org/10.1007/978-3-319-24574-4_28
  9. “Natural language processing (almost) from scratch” In Journal of machine learning research 12.ARTICLE, 2011, pp. 2493–2537
  10. Alex Graves, Abdel-rahman Mohamed and Geoffrey Hinton “Speech recognition with deep recurrent neural networks” In 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, 2013, pp. 6645–6649 DOI: 10.1109/ICASSP.2013.6638947
  11. M. Raissi, P. Perdikaris and G.E. Karniadakis “Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations” In Journal of Computational Physics 378, 2019, pp. 686–707 DOI: https://doi.org/10.1016/j.jcp.2018.10.045
  12. Guofei Pang, Lu Lu and George Em Karniadakis “fPINNs: Fractional Physics-Informed Neural Networks” In SIAM Journal on Scientific Computing 41.4, 2019, pp. A2603–A2626 DOI: https://doi.org/10.1137/18M1229845
  13. “Quantifying total uncertainty in physics-informed neural networks for solving forward and inverse stochastic problems” In Journal of Computational Physics 397, 2019, pp. 108850 DOI: https://doi.org/10.1016/j.jcp.2019.07.048
  14. Ehsan Kharazmi, Zhongqiang Zhang and George E.M. Karniadakis “hp-VPINNs: Variational physics-informed neural networks with domain decomposition” In Computer Methods in Applied Mechanics and Engineering 374 Elsevier BV, 2021, pp. 113547 DOI: 10.1016/j.cma.2020.113547
  15. Stefano Berrone, Claudio Canuto and Moreno Pintore “Variational Physics Informed Neural Networks: the Role of Quadratures and Test Functions” In Journal of Scientific Computing 92.3, 2022, pp. 100 DOI: 10.1007/s10915-022-01950-4
  16. “Exact imposition of boundary conditions with distance functions in physics-informed deep neural networks” In Computer Methods in Applied Mechanics and Engineering 389 Elsevier BV, 2022, pp. 114333 DOI: 10.1016/j.cma.2021.114333
  17. Santiago Badia, Wei Li and Alberto F. Martín “Finite element interpolated neural networks for solving forward and inverse problems” In Computer Methods in Applied Mechanics and Engineering 418, 2024, pp. 116505 DOI: https://doi.org/10.1016/j.cma.2023.116505
  18. “Robust Variational Physics-Informed Neural Networks”, 2023 arXiv:2308.16910 [math.NA]
  19. Sifan Wang, Xinling Yu and Paris Perdikaris “When and why PINNs fail to train: A neural tangent kernel perspective” In Journal of Computational Physics 449, 2022, pp. 110768 DOI: https://doi.org/10.1016/j.jcp.2021.110768
  20. Zhiqiang Cai, Jingshuang Chen and Min Liu “Self-adaptive deep neural network: Numerical approximation to functions and PDEs” In Journal of Computational Physics 455, 2022, pp. 111021 DOI: https://doi.org/10.1016/j.jcp.2022.111021
  21. Ameya D. Jagtap, Kenji Kawaguchi and George Em Karniadakis “Adaptive activation functions accelerate convergence in deep and physics-informed neural networks” In Journal of Computational Physics 404, 2020, pp. 109136 DOI: https://doi.org/10.1016/j.jcp.2019.109136
  22. “DeepXDE: A Deep Learning Library for Solving Differential Equations” In SIAM Review 63.1, 2021, pp. 208–228 DOI: 10.1137/19M1274067
  23. Kejun Tang, Xiaoliang Wan and Chao Yang “DAS-PINNs: A deep adaptive sampling method for solving high-dimensional partial differential equations” In Journal of Computational Physics 476, 2023, pp. 111868 DOI: https://doi.org/10.1016/j.jcp.2022.111868
  24. Jie Hou, Ying Li and Shihui Ying “Enhancing PINNs for solving PDEs via adaptive collocation point movement and adaptive loss weighting” In Nonlinear Dynamics 111.16, 2023, pp. 15233–15261 DOI: https://doi.org/10.1007/s11071-023-08654-w
  25. “Physics-informed neural networks with residual/gradient-based adaptive sampling methods for solving partial differential equations with sharp solutions” In Applied Mathematics and Mechanics 44.7, 2023, pp. 1069–1084 DOI: 10.1007/s10483-023-2994-7
  26. “Adaptive quadratures for nonlinear approximation of low-dimensional PDEs using smooth neural networks”, 2024 DOI: https://doi.org/10.48550/arXiv.2303.11617
  27. “A posteriori error analysis and adaptive processes in the finite element method: Part I—error analysis” In International Journal for Numerical Methods in Engineering 19.11, 1983, pp. 1593–1619 DOI: https://doi.org/10.1002/nme.1620191103
  28. “A priori and a posteriori error estimates for the Deep Ritz method applied to the Laplace and Stokes problem” In Journal of Computational and Applied Mathematics 421 Elsevier BV, 2023, pp. 114845 DOI: 10.1016/j.cam.2022.114845
  29. Stefano Berrone, Claudio Canuto and Moreno Pintore “Solving PDEs by variational physics-informed neural networks: an a posteriori error analysis” In ANNALI DELL’UNIVERSITA’ DI FERRARA 68.2, 2022, pp. 575–595 DOI: 10.1007/s11565-022-00441-6
  30. “A Posteriori Error Control of Approximate Solutions to Boundary Value Problems Found by Neural Networks” In Journal of Mathematical Sciences 273.4 Springer ScienceBusiness Media LLC, 2023, pp. 492–510 DOI: 10.1007/s10958-023-06516-9
  31. “Adaptive Deep Fourier Residual method via overlapping domain decomposition”, 2024 arXiv:2401.04663 [math.NA]
  32. “An analysis of the petrov—galerkin finite element method” In Computer Methods in Applied Mechanics and Engineering 14.1, 1978, pp. 39–64 DOI: https://doi.org/10.1016/0045-7825(78)90012-9
  33. Yeonjong Shin, Zhongqiang Zhang and George Em Karniadakis “Error estimates of residual minimization using neural networks for linear PDEs” In Journal of Machine Learning for Modeling and Computing 4.4 Begell House, 2023, pp. 73–101 DOI: 10.1615/jmachlearnmodelcomput.2023050411
  34. Ignacio Brevis, Ignacio Muga and Kristoffer G. Zee “Neural control of discrete weak formulations: Galerkin, least squares and minimal-residual methods with quasi-optimal weights” In Computer Methods in Applied Mechanics and Engineering 402 Elsevier BV, 2022, pp. 115716 DOI: 10.1016/j.cma.2022.115716
  35. Philipp Petersen, Mones Raslan and Felix Voigtlaender “Topological Properties of the Set of Functions Generated by Neural Networks of Fixed Size” In Foundations of Computational Mathematics 21.2 Springer ScienceBusiness Media LLC, 2020, pp. 375–444 DOI: 10.1007/s10208-020-09461-0
  36. “Preconditioning discretizations of systems of partial differential equations” In Numerical Linear Algebra with Applications 18.1 Wiley, 2010, pp. 1–40 DOI: 10.1002/nla.716
  37. “Variational Multiscale Methods in Computational Fluid Dynamics” In Encyclopedia of Computational Mechanics Second Edition Wiley, 2017, pp. 1–28 DOI: 10.1002/9781119176817.ecm2117
  38. Santiago Badia “On stabilized finite element methods based on the Scott–Zhang projector. Circumventing the inf–sup condition for the Stokes problem” In Computer Methods in Applied Mechanics and Engineering 247–248 Elsevier BV, 2012, pp. 65–72 DOI: 10.1016/j.cma.2012.07.020
  39. Carsten Burstedde, Lucas C. Wilcox and Omar Ghattas “p4est: Scalable Algorithms for Parallel Adaptive Mesh Refinement on Forests of Octrees” In SIAM Journal on Scientific Computing 33.3, 2011, pp. 1103–1133 DOI: 10.1137/100791634
  40. “A Generic Finite Element Framework on Parallel Tree-Based Adaptive Meshes” In SIAM Journal on Scientific Computing 42.6, 2020, pp. C436–C468 DOI: 10.1137/20M1328786
  41. Alberto F. Martín “GridapP4est.jl”, 2024 URL: https://github.com/gridap/GridapP4est.jl
  42. Marc Olm, Santiago Badia and Alberto F. Martín “On a general implementation of h- and p-adaptive curl-conforming finite elements” In Advances in Engineering Software 132 Elsevier BV, 2019, pp. 74–91 DOI: 10.1016/j.advengsoft.2019.03.006
  43. “Gridap: An extensible Finite Element toolbox in Julia” In Journal of Open Source Software 5.52 The Open Journal, 2020, pp. 2520 DOI: 10.21105/joss.02520
  44. “The software design of Gridap: A Finite Element package based on the Julia JIT compiler” In Computer Physics Communications 276 Elsevier BV, 2022, pp. 108341 DOI: 10.1016/j.cpc.2022.108341
  45. “Fashionable Modelling with Flux” In CoRR abs/1811.01457, 2018 arXiv: https://arxiv.org/abs/1811.01457
  46. Mike Innes “Flux: Elegant Machine Learning with Julia” In Journal of Open Source Software, 2018 DOI: 10.21105/joss.00602
  47. Michael Innes “Don’t Unroll Adjoint: Differentiating SSA-Form Programs” In CoRR abs/1810.07951, 2018 arXiv: http://arxiv.org/abs/1810.07951
  48. “Understanding the difficulty of training deep feedforward neural networks” In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics 9, Proceedings of Machine Learning Research Chia Laguna Resort, Sardinia, Italy: PMLR, 2010, pp. 249–256 URL: https://proceedings.mlr.press/v9/glorot10a.html
  49. Patrick Kofod Mogensen and Asbjørn Nilsen Riseth “Optim: A mathematical optimization package for Julia” In Journal of Open Source Software 3.24, 2018, pp. 615 DOI: 10.21105/joss.00615
  50. “Gmsh: A 3-D finite element mesh generator with built-in pre- and post-processing facilities” In International Journal for Numerical Methods in Engineering 79.11, 2009, pp. 1309–1331 DOI: https://doi.org/10.1002/nme.2579
  51. Marc Olm, Santiago Badia and Alberto F. Martín “On a general implementation of h- and p-adaptive curl-conforming finite elements” In Advances in Engineering Software 132, 2019, pp. 74–91 DOI: https://doi.org/10.1016/j.advengsoft.2019.03.006
  52. “The Aggregated Unfitted Finite Element Method on Parallel Tree-Based Adaptive Meshes” In SIAM Journal on Scientific Computing 43.3 Society for Industrial & Applied Mathematics (SIAM), 2021, pp. C203–C234 DOI: 10.1137/20m1344512
Citations (1)

Summary

We haven't generated a summary for this paper yet.