Further Understanding of a Local Gaussian Process Approximation: Characterising Convergence in the Finite Regime (2404.06200v1)
Abstract: We show that common choices of kernel functions for a highly accurate and massively scalable nearest-neighbour based GP regression model (GPnn: \cite{GPnn}) exhibit gradual convergence to asymptotic behaviour as dataset-size $n$ increases. For isotropic kernels such as Mat\'{e}rn and squared-exponential, an upper bound on the predictive MSE can be obtained as $O(n{-\frac{p}{d}})$ for input dimension $d$, $p$ dictated by the kernel (and $d>p$) and fixed number of nearest-neighbours $m$ with minimal assumptions on the input distribution. Similar bounds can be found under model misspecification and combined to give overall rates of convergence of both MSE and an important calibration metric. We show that lower bounds on $n$ can be given in terms of $m$, $l$, $p$, $d$, a tolerance $\varepsilon$ and a probability $\delta$. When $m$ is chosen to be $O(n{\frac{p}{p+d}})$ minimax optimal rates of convergence are attained. Finally, we demonstrate empirical performance and show that in many cases convergence occurs faster than the upper bounds given here.
- “Leveraging Locality and Robustness to Achieve Massively Scalable Gaussian Process Regression” In Advances in Neural Information Processing Systems 36 Curran Associates, Inc., 2023 arXiv:2306.14731 [stat.ML]
- David R Burt and Carl Edward Rasmussen “Convergence of Sparse Variational Inference in Gaussian Processes Regression” In Journal of Machine Learning Research 21, 2020, pp. 1–63
- “Hierarchical Nearest-Neighbor Gaussian Process Models for Large Geostatistical Datasets” In Journal of the American Statistical Association 111.514, 2016, pp. 800–812 DOI: 10.1080/01621459.2015.1044091
- Eleonora Di Nezza, Giampiero Palatucci and Enrico Valdinoci “Hitchhiker’s guide to the fractional Sobolev spaces” In Bulletin des Sciences Mathématiques 136.5, 2012, pp. 521–573 DOI: https://doi.org/10.1016/j.bulsci.2011.12.004
- “NIST Digital Library of Mathematical Functions” F. W. J. Olver, A. B. Olde Daalhuis, D. W. Lozier, B. I. Schneider, R. F. Boisvert, C. W. Clark, B. R. Miller, B. V. Saunders, H. S. Cohl, and M. A. McClain, eds., https://dlmf.nist.gov/, Release 1.1.11 of 2023-09-15 URL: https://dlmf.nist.gov/
- European Mathematical Society “Neumann series” In Encyclopedia of Mathematics Springer, http://encyclopediaofmath.org/index.php?title=Neumann_series&oldid=55168, 2024
- “Efficient algorithms for Bayesian Nearest Neighbor Gaussian Processes” arXiv:1702.00434 [stat] arXiv, 2018 URL: http://arxiv.org/abs/1702.00434
- “Fitting Mat\’ern Smoothness Parameters Using Automatic Differentiation” arXiv: 2201.00090, 2021 URL: http://arxiv.org/abs/2201.00090
- Robert B. Gramacy and Daniel W. Apley “Local Gaussian Process Approximation for Large Computer Experiments” arXiv: 1303.0383 In Journal of Computational and Graphical Statistics 24.2, 2015, pp. 561–578 DOI: 10.1080/10618600.2014.914442
- “A Distribution-Free Theory of Nonparametric Regression” Springer Series in Statistics, 2010
- Roger A. Horn and Charles R. Johnson “Matrix Analysis” In Matrix Analysis Cambridge University Press, 1985 DOI: 10.1017/cbo9780511810817
- Michael Kohler, Adam Krzyzak and Harro Walk “Rates of convergence for partitioning and nearest neighbor regression estimates with unbounded data” In Journal of Multivariate Analysis 97.2, 2006, pp. 311–323 DOI: 10.1016/j.jmva.2005.03.006
- “When Gaussian process meets big data: A review of scalable GPs” In IEEE transactions on neural networks and learning systems 31.11 IEEE, 2020, pp. 4405–4423
- Dennis Nieman, Botond Szabo and Harry Zanten “Contraction rates for sparse variational approximations in Gaussian process regression” In Journal of Machine Learning Research 23.205, 2022, pp. 1–26 URL: http://jmlr.org/papers/v23/21-1128.html
- Bernhard Schölkopf “The Kernel Trick for Distances” In Advances in Neural Information Processing Systems 13 MIT Press, 2000
- Michael L. Stein “2010 Rietz lecture when does the screening effect hold?” arXiv: 1203.1801v1 In Annals of Statistics 39.6, 2011, pp. 2795–2819 DOI: 10.1214/11-AOS909
- Michael L. Stein “The screening effect in kriging” In Annals of Statistics 30.1, 2002, pp. 298–323 DOI: 10.1214/aos/1015362194
- Michael L. Stein, Zhiyi Chi and Leah J. Welty “Approximating likelihoods for large spatial data sets” In Journal of the Royal Statistical Society. Series B: Statistical Methodology 66.2, 2004, pp. 275–296 DOI: 10.1046/j.1369-7412.2003.05512.x
- “Sparse within Sparse Gaussian Processes using Neighbor Information” In Proceedings of the 38th International Conference on Machine Learning 139, Proceedings of Machine Learning Research PMLR, 2021, pp. 10369–10378 URL: https://proceedings.mlr.press/v139/tran21a.html
- Filip Tronarp, Toni Karvonen and Simo Sarkka “Mixture representation of the matérn class with applications in state space approximations and Bayesian quadrature” In IEEE International Workshop on Machine Learning for Signal Processing, MLSP 2018-September.2, 2018 DOI: 10.1109/MLSP.2018.8516992
- Alexandre B. Tsybakov “Introduction to Nonparametric Estimation” Springer Publishing Company, Incorporated, 2008
- Aad Vaart “Information Rates of Nonparametric Gaussian Process Methods” In Journal of Machine Learning Research 12, 2011 URL: https://jmlr.org/papers/volume12/vandervaart11a/vandervaart11a.pdf
- “Improved Convergence Rates for Sparse Approximation Methods in Kernel-Based Learning” ISSN: 2640-3498 In Proceedings of the 39th International Conference on Machine Learning PMLR, 2022, pp. 21960–21983 URL: https://proceedings.mlr.press/v162/vakili22a.html
- A.V. Vecchia “Estimation and Model Identification for Continuous Spatial Processes” Publisher: [Royal Statistical Society, Wiley] In Journal of the Royal Statistical Society. Series B (Methodological) 50.2, 1988, pp. 297–312 URL: https://www.jstor.org/stable/2345768
- Christopher K Williams and Carl Edward Rasmussen “Gaussian processes for machine learning” MIT press Cambridge, MA, 2006
- Luhuan Wu, Geoff Pleiss and John Cunningham “Variational Nearest Neighbor Gaussian Process” In Proceedings of the 39th International Conference on Machine Learning, 2022 arXiv: http://arxiv.org/abs/2202.01694
Sponsor
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.