Learning from Linear Algebra: A Graph Neural Network Approach to Preconditioner Design for Conjugate Gradient Solvers (2405.15557v2)
Abstract: Large linear systems are ubiquitous in modern computational science and engineering. The main recipe for solving them is the use of Krylov subspace iterative methods with well-designed preconditioners. Deep learning models can be used as nonlinear preconditioners during the iteration of linear solvers such as the conjugate gradient (CG) method. Neural network models require an enormous number of parameters to approximate well in this setup. Another approach is to take advantage of small graph neural networks (GNNs) to construct preconditioners with predefined sparsity patterns. Recently, GNNs have been shown to be a promising tool for designing preconditioners to reduce the overall computational cost of iterative methods by constructing them more efficiently than with classical linear algebra techniques. However, preconditioners designed with these approaches cannot outperform those designed with classical methods in terms of the number of iterations in CG. In our work, we recall well-established preconditioners from linear algebra and use them as a starting point for training the GNN to obtain preconditioners that reduce the condition number of the system more significantly. Numerical experiments show that our approach outperforms both classical and neural network-based methods for an important class of parametric partial differential equations. We also provide a heuristic justification for the loss function used and show that preconditioners obtained by learning with this loss function reduce the condition number in a more desirable way for CG.
- Yousef Saad. Iterative methods for sparse linear systems. SIAM, 2003.
- Neural operators meet conjugate gradients: The fcg-no method for efficient pde solving. arXiv preprint arXiv:2402.05598, 2024.
- Maksym Shpakovych. Neural network preconditioning of large linear systems. PhD thesis, Inria Centre at the University of Bordeaux, 2023.
- Deeponet based preconditioning strategies for solving parametric linear systems of equations. arXiv preprint arXiv:2401.02016, 2024.
- Fourier neural solver for large sparse linear algebraic systems. Mathematics, 10(21):4014, 2022.
- Learning preconditioners for conjugate gradient pde solvers. In International Conference on Machine Learning, pages 19425–19439. PMLR, 2023.
- Neural incomplete factorization: learning preconditioners for the conjugate gradient method. arXiv preprint arXiv:2305.16368, 2023.
- Numerical linear algebra. SIAM, 2022.
- Michael F Hutchinson. A stochastic estimator of the trace of the influence matrix for laplacian smoothing splines. Communications in Statistics-Simulation and Computation, 18(3):1059–1076, 1989.
- Graph neural networks: A review of methods and applications. AI open, 1:57–81, 2020.
- Message passing neural pde solvers. arXiv preprint arXiv:2202.03376, 2022.
- EJ Carr and IW Turner. A semi-analytical solution for multilayer diffusion in a composite medium consisting of a large number of layers. Applied Mathematical Modelling, 40(15-16):7034–7050, 2016.
- Diffusion of electromagnetic fields into a two-dimensional earth: A finite-difference approach. Geophysics, 49(7):870–894, 1984.
- Multigrid pressure solver for 2d displacement problems in drilling, cementing, fracturing and eor. Journal of Petroleum Science and Engineering, 196:107918, 2021.
- Jan Mayer. A multilevel crout ilu preconditioner with pivoting and row permutation. Numerical Linear Algebra with Applications, 14(10):771–789, 2007.
- Owe Axelsson. Iterative solution methods. Cambridge university press, 1996.
- A hybrid iterative numerical transferable solver (hints) for pdes based on deep operator network and relaxation methods. arXiv preprint arXiv:2208.13273, 2022.