Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Dropout Prediction Uncertainty Estimation Using Neuron Activation Strength (2110.06435v3)

Published 13 Oct 2021 in cs.LG

Abstract: Dropout has been commonly used to quantify prediction uncertainty, i.e, the variations of model predictions on a given input example. However, using dropout in practice can be expensive as it requires running dropout inferences many times. In this paper, we study how to estimate dropout prediction uncertainty in a resource-efficient manner. We demonstrate that we can use neuron activation strengths to estimate dropout prediction uncertainty under different dropout settings and on a variety of tasks using three large datasets, MovieLens, Criteo, and EMNIST. Our approach provides an inference-once method to estimate dropout prediction uncertainty as a cheap auxiliary task. We also demonstrate that using activation features from a subset of the neural network layers can be sufficient to achieve uncertainty estimation performance almost comparable to that of using activation features from all layers, thus reducing resources even further for uncertainty estimation.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Haichao Yu (11 papers)
  2. Zhe Chen (237 papers)
  3. Dong Lin (15 papers)
  4. Gil Shamir (4 papers)
  5. Jie Han (93 papers)

Summary

We haven't generated a summary for this paper yet.