A Review of Uncertainty Quantification in Deep Learning: Techniques, Applications and Challenges
Published 12 Nov 2020 in cs.LG, cs.AI, and cs.CV | (2011.06225v4)
Abstract: Uncertainty quantification (UQ) plays a pivotal role in reduction of uncertainties during both optimization and decision making processes. It can be applied to solve a variety of real-world applications in science and engineering. Bayesian approximation and ensemble learning techniques are two most widely-used UQ methods in the literature. In this regard, researchers have proposed different UQ methods and examined their performance in a variety of applications such as computer vision (e.g., self-driving cars and object detection), image processing (e.g., image restoration), medical image analysis (e.g., medical image classification and segmentation), natural language processing (e.g., text classification, social media texts and recidivism risk-scoring), bioinformatics, etc. This study reviews recent advances in UQ methods used in deep learning. Moreover, we also investigate the application of these methods in reinforcement learning (RL). Then, we outline a few important applications of UQ methods. Finally, we briefly highlight the fundamental research challenges faced by UQ methods and discuss the future research directions in this field.
The paper reviews key uncertainty quantification techniques in deep learning, including Bayesian approaches and ensemble methods.
The paper contrasts aleatoric and epistemic uncertainties to highlight practical implications for enhancing AI system reliability.
The paper outlines current challenges and future research directions, emphasizing methods for integrating uncertainty in reinforcement learning.
The paper "A Review of Uncertainty Quantification in Deep Learning: Techniques, Applications and Challenges" presents a comprehensive survey of recent methods for uncertainty quantification (UQ) in deep learning (DL). The authors discuss various UQ techniques, including Bayesian approximation and ensemble learning, and their applications across diverse fields. The study also highlights research challenges and future directions within the UQ domain.
The paper begins by defining the two primary types of uncertainty: aleatoric and epistemic. Aleatoric uncertainty, also known as data uncertainty, arises from irreducible noise in the data. Epistemic uncertainty, also known as knowledge uncertainty, stems from a lack of knowledge or data. The authors emphasize the importance of quantifying these uncertainties to improve the trustworthiness and accuracy of AI systems.
The paper then provides a detailed review of UQ methods based on Bayesian techniques:
Bayesian Neural Networks (BNNs) are introduced as a way to represent model parameters as probability distributions, offering robustness against overfitting.
Monte Carlo Dropout (MCD) is presented as an efficient approximation to Bayesian inference, using dropout as a regularization technique to estimate prediction uncertainty. The loss function with L2 regularization is expressed as:
E(yi,yi^) is the error between the true value yi and the predicted value yi^
λ is the regularization parameter
L is the number of layers in the neural network
Wi is the weight matrix for layer i
bi is the bias vector for layer i
Markov Chain Monte Carlo (MCMC) methods are discussed for approximating posterior distributions, with a focus on Stochastic Gradient MCMC (SG-MCMC) for training DNNs.
Variational Inference (VI) is presented as an optimization-based approach to approximate posterior distributions in BNNs. The loss is defined as:
KL(qϕ(w)∥p(w)) is the Kullback-Leibler divergence between the approximate posterior qϕ(w) and the prior p(w)
Bayesian Active Learning (BAL) is introduced as a method to select the most informative unlabeled samples for annotation, improving learning efficiency.
Bayes by Backprop (BBB) is described as an algorithm for quantifying uncertainty in neural network weights by learning a probability distribution over the weights.
Variational Autoencoders (VAEs) are presented as generative models for learning representations and modeling posterior distributions.
The study also reviews other UQ methods beyond Bayesian techniques:
Deep Gaussian Processes (DGPs) are discussed as multi-layer models for accurate uncertainty modeling.
The application of UQ in Reinforcement Learning (RL) is explored, where uncertainty plays a critical role in decision-making. The paper highlights the use of Bayesian methods and ensemble techniques for quantifying uncertainty in RL agents. The authors discuss Bayesian Policy Optimization (BPO) method applied to Partially Observable Markov Decision Processes (POMDPs) as a Bayes filter to compute the belief b of the hidden state as follows:
b′(s′)=ψ(b,a′,o′)=ηs∈S∑b(s)T(s,a′,s′)Z(s,a′,o′)
where:
b′ is the updated belief state
b is the previous belief state
a′ is the action
o′ is the observation
η is a normalization factor
S is the state space
T is the transition function
Z is the observation function
Ensemble techniques are also discussed as a means to improve predictive performance and quantify uncertainty. The authors describe how ensembles can capture different sources of model uncertainty and provide more reliable estimates. The total uncertainty can be decomposed into expected data uncertainty and knowledge uncertainty via Mutual Information (MI) formulation as follows:
The paper discusses various ensemble methods such as deep ensembles, deep ensemble Bayesian, and UQ in traditional machine learning domain using ensemble techniques, providing an overview of their strengths and weaknesses.
The study includes a comprehensive review of recent articles published in quantifying uncertainty in AI (ML and DL) using different approaches. The review identifies main research gaps of UQ methods and pointing out solid future directions for researchers in this domain.