Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Scalable Hyperparameter Optimization with Lazy Gaussian Processes (2001.05726v1)

Published 16 Jan 2020 in cs.LG and stat.ML

Abstract: Most machine learning methods require careful selection of hyper-parameters in order to train a high performing model with good generalization abilities. Hence, several automatic selection algorithms have been introduced to overcome tedious manual (try and error) tuning of these parameters. Due to its very high sample efficiency, Bayesian Optimization over a Gaussian Processes modeling of the parameter space has become the method of choice. Unfortunately, this approach suffers from a cubic compute complexity due to underlying Cholesky factorization, which makes it very hard to be scaled beyond a small number of sampling steps. In this paper, we present a novel, highly accurate approximation of the underlying Gaussian Process. Reducing its computational complexity from cubic to quadratic allows an efficient strong scaling of Bayesian Optimization while outperforming the previous approach regarding optimization accuracy. The first experiments show speedups of a factor of 162 in single node and further speed up by a factor of 5 in a parallel environment.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Raju Ram (1 paper)
  2. Sabine Müller (7 papers)
  3. Franz-Josef Pfreundt (22 papers)
  4. Nicolas R. Gauger (41 papers)
  5. Janis Keuper (66 papers)
Citations (5)

Summary

We haven't generated a summary for this paper yet.