Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Weighting Is Worth the Wait: Bayesian Optimization with Importance Sampling (2002.09927v1)

Published 23 Feb 2020 in cs.LG and stat.ML

Abstract: Many contemporary machine learning models require extensive tuning of hyperparameters to perform well. A variety of methods, such as Bayesian optimization, have been developed to automate and expedite this process. However, tuning remains extremely costly as it typically requires repeatedly fully training models. We propose to accelerate the Bayesian optimization approach to hyperparameter tuning for neural networks by taking into account the relative amount of information contributed by each training example. To do so, we leverage importance sampling (IS); this significantly increases the quality of the black-box function evaluations, but also their runtime, and so must be done carefully. Casting hyperparameter search as a multi-task Bayesian optimization problem over both hyperparameters and importance sampling design achieves the best of both worlds: by learning a parameterization of IS that trades-off evaluation complexity and quality, we improve upon Bayesian optimization state-of-the-art runtime and final validation error across a variety of datasets and complex neural architectures.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Setareh Ariafar (3 papers)
  2. Zelda Mariet (15 papers)
  3. Ehsan Elhamifar (13 papers)
  4. Dana Brooks (3 papers)
  5. Jennifer Dy (46 papers)
  6. Jasper Snoek (42 papers)
Citations (3)

Summary

We haven't generated a summary for this paper yet.