Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
156 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Sparse Variational Contaminated Noise Gaussian Process Regression with Applications in Geomagnetic Perturbations Forecasting (2402.17570v3)

Published 27 Feb 2024 in cs.LG, stat.AP, and stat.ME

Abstract: Gaussian Processes (GP) have become popular machine-learning methods for kernel-based learning on datasets with complicated covariance structures. In this paper, we present a novel extension to the GP framework using a contaminated normal likelihood function to better account for heteroscedastic variance and outlier noise. We propose a scalable inference algorithm based on the Sparse Variational Gaussian Process (SVGP) method for fitting sparse Gaussian process regression models with contaminated normal noise on large datasets. We examine an application to geomagnetic ground perturbations, where the state-of-the-art prediction model is based on neural networks. We show that our approach yields shorter prediction intervals for similar coverage and accuracy when compared to an artificial dense neural network baseline.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (40)
  1. Jesper W. Gjerloev. The SuperMAG data processing technique. Journal of Geophysical Research, 117, 2012.
  2. Assessing the impact of space weather on the electric power grid based on insurance claims for industrial electrical equipment. Space Weather, 12(7):487–498, 2014. _eprint: https://agupubs.onlinelibrary.wiley.com/doi/pdf/10.1002/2014SW001066.
  3. OMNI 1-min Data. NASA Space Physics Data Facility, 2020. https://doi.org/10.48322/45bb-8792, Accessed on Sept. 1, 2021.
  4. John R. Gleason. Understanding Elongation: The Scale Contaminated Normal Family. Journal of the American Statistical Association, 88(421):327–337, 1993. Publisher: [American Statistical Association, Taylor & Francis, Ltd.].
  5. A Bayesian Approach to Some Outlier Problems. Biometrika, 55(1):119–129, 1968. Publisher: [Oxford University Press, Biometrika Trust].
  6. A Variational Approach to Robust Regression. In Georg Dorffner, Horst Bischof, and Kurt Hornik, editors, Artificial Neural Networks — ICANN 2001, pages 95–102, Berlin, Heidelberg, 2001. Springer Berlin Heidelberg.
  7. DeepGUM: Learning Deep Robust Regression with a Gaussian-Uniform Mixture Model. In Vittorio Ferrari, Martial Hebert, Cristian Sminchisescu, and Yair Weiss, editors, Computer Vision – ECCV 2018, volume 11209, pages 205–221. Springer International Publishing, Cham, 2018. Series Title: Lecture Notes in Computer Science.
  8. Robust sparse regression by modeling noise as a mixture of gaussians. Journal of Applied Statistics, 46(10):1738–1755, July 2019.
  9. Robust probabilistic principal component regression with switching mixture Gaussian noise for soft sensing. Chemometrics and Intelligent Laboratory Systems, 222:104491, 2022.
  10. M. Kuss. Gaussian Process Models for Robust Regression, Classification, and Reinforcement Learning. PhD thesis, Technische Universität Darmstadt, 2006.
  11. Gaussian process modelling with Gaussian mixture likelihood. Journal of Process Control, 81:209–220, September 2019.
  12. Andrew Naish and S. Holden. Robust Regression with Twinned Gaussian Processes. In NIPS, 2007.
  13. Robust Gaussian Process Regression with a Student- t Likelihood. Journal of Machine Learning Research, 12, June 2011.
  14. Robust Gaussian Process Regression Based on Iterative Trimming. Astronomy and Computing, 36:100483, July 2021. arXiv:2011.11057 [astro-ph, stat].
  15. Robust Gaussian Process Regression with Huber Likelihood, January 2023. Number: arXiv:2301.07858 arXiv:2301.07858 [stat].
  16. When Gaussian Process Meets Big Data: A Review of Scalable GPs, April 2019. Number: arXiv:1807.01065 arXiv:1807.01065 [cs, stat].
  17. A Unifying View of Sparse Approximate Gaussian Process Regression. Journal of Machine Learning Research, 6(65):1939–1959, 2005.
  18. Fast Forward Selection to Speed Up Sparse Gaussian Process Regression. In Christopher M. Bishop and Brendan J. Frey, editors, Proceedings of the Ninth International Workshop on Artificial Intelligence and Statistics, volume R4 of Proceedings of Machine Learning Research, pages 254–261. PMLR, January 2003.
  19. Sparse Gaussian Processes using Pseudo-inputs. In Y. Weiss, B. Schölkopf, and J. Platt, editors, Advances in Neural Information Processing Systems, volume 18. MIT Press, 2005.
  20. Local and global sparse Gaussian process approximations. Journal of Machine Learning Research - Proceedings Track, 2:524–531, 2007.
  21. Inter-domain Gaussian Processes for Sparse Inference using Inducing Features. In Y. Bengio, D. Schuurmans, J. Lafferty, C. Williams, and A. Culotta, editors, Advances in Neural Information Processing Systems, volume 22. Curran Associates, Inc., 2009.
  22. Michalis Titsias. Variational Learning of Inducing Variables in Sparse Gaussian Processes. In David van Dyk and Max Welling, editors, Proceedings of the Twelth International Conference on Artificial Intelligence and Statistics, volume 5 of Proceedings of Machine Learning Research, pages 567–574, Hilton Clearwater Beach Resort, Clearwater Beach, Florida USA, April 2009. PMLR.
  23. Gaussian Processes for Big Data. Technical Report arXiv:1309.6835, arXiv, September 2013. arXiv:1309.6835 [cs, stat] type: article.
  24. Scalable Variational Gaussian Process Classification. In Guy Lebanon and S. V. N. Vishwanathan, editors, Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics, volume 38 of Proceedings of Machine Learning Research, pages 351–360. PMLR, May 2015.
  25. GPyTorch: Blackbox Matrix-Matrix Gaussian Process Inference with GPU Acceleration. In Neural Information Processing Systems, 2018.
  26. Carl Edward Rasmussen and Christopher K. I. Williams. Gaussian Processes for Machine Learning (Adaptive Computation and Machine Learning). The MIT Press, 2005.
  27. Understanding predictive information criteria for Bayesian models. Statistics and Computing, 24(6):997–1016, November 2014.
  28. A Note on Gauss-Hermite Quadrature. Biometrika, 81(3):624–629, 1994. Publisher: [Oxford University Press, Biometrika Trust].
  29. Adam: A method for stochastic optimization, 2014.
  30. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Proceedings of the 33rd International Conference on Neural Information Processing Systems. Curran Associates Inc., Red Hook, NY, USA, 2019.
  31. Radford M. Neal. Monte Carlo Implementation of Gaussian Process Models for Bayesian Regression and Classification. arXiv: Data Analysis, Statistics and Probability, 1997.
  32. Jerome H. Friedman. Multivariate Adaptive Regression Splines. The Annals of Statistics, 19(1), March 1991.
  33. Fernando Nogueira. Bayesian Optimization: Open source constrained global optimization tool for Python, 2014–.
  34. Comparison of Deep Learning Techniques to Model Connections Between Solar Wind and Ground Magnetic Perturbations. Frontiers in Astronomy and Space Sciences, 7:550874, October 2020.
  35. Global Geomagnetic Perturbation Forecasting Using Deep Learning. Space Weather, 20(6), June 2022.
  36. Revisiting the Ground Magnetic Field Perturbations Challenge: A Machine Learning Perspective. Frontiers in Astronomy and Space Sciences, 9:869740, May 2022.
  37. A Gray-Box Model for a Probabilistic Estimate of Regional Ground Magnetic Perturbations: Enhancing the NOAA Operational Geospace Model With Machine Learning. J. Geophys. Res. Space Physics, 125(11), 2020.
  38. Variational Inference: A Review for Statisticians. Journal of the American Statistical Association, 112(518):859–877, April 2017. Publisher: Taylor & Francis.
  39. Graphical Models, Exponential Families, and Variational Inference. Foundations and Trends® in Machine Learning, 1(1–2):1–305, 2007.
  40. The Variational Gaussian Approximation Revisited. Neural Computation, 21(3):786–792, March 2009.
Citations (1)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com