Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Efficient hyperparameter optimization by way of PAC-Bayes bound minimization (2008.06431v1)

Published 14 Aug 2020 in stat.ML and cs.LG

Abstract: Identifying optimal values for a high-dimensional set of hyperparameters is a problem that has received growing attention given its importance to large-scale machine learning applications such as neural architecture search. Recently developed optimization methods can be used to select thousands or even millions of hyperparameters. Such methods often yield overfit models, however, leading to poor performance on unseen data. We argue that this overfitting results from using the standard hyperparameter optimization objective function. Here we present an alternative objective that is equivalent to a Probably Approximately Correct-Bayes (PAC-Bayes) bound on the expected out-of-sample error. We then devise an efficient gradient-based algorithm to minimize this objective; the proposed method has asymptotic space and time complexity equal to or better than other gradient-based hyperparameter optimization methods. We show that this new method significantly reduces out-of-sample error when applied to hyperparameter optimization problems known to be prone to overfitting.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. John J. Cherian (4 papers)
  2. Andrew G. Taube (1 paper)
  3. Robert T. McGibbon (7 papers)
  4. Panagiotis Angelikopoulos (5 papers)
  5. Guy Blanc (30 papers)
  6. Michael Snarski (6 papers)
  7. Daniel D. Richman (1 paper)
  8. John L. Klepeis (1 paper)
  9. David E. Shaw (5 papers)
Citations (5)

Summary

We haven't generated a summary for this paper yet.