Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

The Hidden Uncertainty in a Neural Networks Activations (2012.03082v2)

Published 5 Dec 2020 in cs.LG and stat.ML

Abstract: The distribution of a neural network's latent representations has been successfully used to detect out-of-distribution (OOD) data. This work investigates whether this distribution moreover correlates with a model's epistemic uncertainty, thus indicates its ability to generalise to novel inputs. We first empirically verify that epistemic uncertainty can be identified with the surprise, thus the negative log-likelihood, of observing a particular latent representation. Moreover, we demonstrate that the output-conditional distribution of hidden representations also allows quantifying aleatoric uncertainty via the entropy of the predictive distribution. We analyse epistemic and aleatoric uncertainty inferred from the representations of different layers and conclude that deeper layers lead to uncertainty with similar behaviour as established - but computationally more expensive - methods (e.g. deep ensembles). While our approach does not require modifying the training process, we follow prior work and experiment with an additional regularising loss that increases the information in the latent representations. We find that this leads to improved OOD detection of epistemic uncertainty at the cost of ambiguous calibration close to the data distribution. We verify our findings on both classification and regression models.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Janis Postels (10 papers)
  2. Hermann Blum (36 papers)
  3. Yannick Strümpler (4 papers)
  4. Cesar Cadena (94 papers)
  5. Roland Siegwart (236 papers)
  6. Luc Van Gool (570 papers)
  7. Federico Tombari (214 papers)
Citations (20)

Summary

We haven't generated a summary for this paper yet.