Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Multi-headed Neural Ensemble Search (2107.04369v1)

Published 9 Jul 2021 in cs.LG and stat.ML

Abstract: Ensembles of CNN models trained with different seeds (also known as Deep Ensembles) are known to achieve superior performance over a single copy of the CNN. Neural Ensemble Search (NES) can further boost performance by adding architectural diversity. However, the scope of NES remains prohibitive under limited computational resources. In this work, we extend NES to multi-headed ensembles, which consist of a shared backbone attached to multiple prediction heads. Unlike Deep Ensembles, these multi-headed ensembles can be trained end to end, which enables us to leverage one-shot NAS methods to optimize an ensemble objective. With extensive empirical evaluations, we demonstrate that multi-headed ensemble search finds robust ensembles 3 times faster, while having comparable performance to other ensemble search methods, in both predictive performance and uncertainty calibration.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Ashwin Raaghav Narayanan (1 paper)
  2. Arber Zela (22 papers)
  3. Tonmoy Saikia (8 papers)
  4. Thomas Brox (134 papers)
  5. Frank Hutter (177 papers)
Citations (4)

Summary

We haven't generated a summary for this paper yet.