Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A general framework for ensemble distribution distillation (2002.11531v2)

Published 26 Feb 2020 in stat.ML and cs.LG

Abstract: Ensembles of neural networks have been shown to give better performance than single networks, both in terms of predictions and uncertainty estimation. Additionally, ensembles allow the uncertainty to be decomposed into aleatoric (data) and epistemic (model) components, giving a more complete picture of the predictive uncertainty. Ensemble distillation is the process of compressing an ensemble into a single model, often resulting in a leaner model that still outperforms the individual ensemble members. Unfortunately, standard distillation erases the natural uncertainty decomposition of the ensemble. We present a general framework for distilling both regression and classification ensembles in a way that preserves the decomposition. We demonstrate the desired behaviour of our framework and show that its predictive performance is on par with standard distillation.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Jakob Lindqvist (5 papers)
  2. Amanda Olmin (5 papers)
  3. Fredrik Lindsten (69 papers)
  4. Lennart Svensson (81 papers)
Citations (18)

Summary

We haven't generated a summary for this paper yet.