The paper "A Review of Uncertainty Quantification in Deep Learning: Techniques, Applications and Challenges" presents a comprehensive survey of recent methods for uncertainty quantification (UQ) in deep learning (DL). The authors discuss various UQ techniques, including Bayesian approximation and ensemble learning, and their applications across diverse fields. The paper also highlights research challenges and future directions within the UQ domain.
The paper begins by defining the two primary types of uncertainty: aleatoric and epistemic. Aleatoric uncertainty, also known as data uncertainty, arises from irreducible noise in the data. Epistemic uncertainty, also known as knowledge uncertainty, stems from a lack of knowledge or data. The authors emphasize the importance of quantifying these uncertainties to improve the trustworthiness and accuracy of AI systems.
The paper then provides a detailed review of UQ methods based on Bayesian techniques:
- Bayesian Neural Networks (BNNs) are introduced as a way to represent model parameters as probability distributions, offering robustness against overfitting.
- Monte Carlo Dropout (MCD) is presented as an efficient approximation to Bayesian inference, using dropout as a regularization technique to estimate prediction uncertainty. The loss function with regularization is expressed as:
where:
- is the number of samples
- is the error between the true value and the predicted value
- is the regularization parameter
- is the number of layers in the neural network
- is the weight matrix for layer
- is the bias vector for layer
- Markov Chain Monte Carlo (MCMC) methods are discussed for approximating posterior distributions, with a focus on Stochastic Gradient MCMC (SG-MCMC) for training DNNs.
- Variational Inference (VI) is presented as an optimization-based approach to approximate posterior distributions in BNNs. The loss is defined as:
where:
- is the number of samples
- is the reconstruction loss
- is the Kullback-Leibler divergence between the approximate posterior and the prior
- Bayesian Active Learning (BAL) is introduced as a method to select the most informative unlabeled samples for annotation, improving learning efficiency.
- Bayes by Backprop (BBB) is described as an algorithm for quantifying uncertainty in neural network weights by learning a probability distribution over the weights.
- Variational Autoencoders (VAEs) are presented as generative models for learning representations and modeling posterior distributions.
The paper also reviews other UQ methods beyond Bayesian techniques:
- Deep Gaussian Processes (DGPs) are discussed as multi-layer models for accurate uncertainty modeling.
- Laplace Approximations (LAs) are presented as a way to approximate Bayesian inference by building a Gaussian distribution around the maximum a posteriori (MAP) estimate.
The application of UQ in Reinforcement Learning (RL) is explored, where uncertainty plays a critical role in decision-making. The paper highlights the use of Bayesian methods and ensemble techniques for quantifying uncertainty in RL agents. The authors discuss Bayesian Policy Optimization (BPO) method applied to Partially Observable Markov Decision Processes (POMDPs) as a Bayes filter to compute the belief of the hidden state as follows:
where:
- is the updated belief state
- is the previous belief state
- is the action
- is the observation
- is a normalization factor
- is the state space
- is the transition function
- is the observation function
Ensemble techniques are also discussed as a means to improve predictive performance and quantify uncertainty. The authors describe how ensembles can capture different sources of model uncertainty and provide more reliable estimates. The total uncertainty can be decomposed into expected data uncertainty and knowledge uncertainty via Mutual Information (MI) formulation as follows:
Where:
- is the predicted output
- is the model parameter
- is the test data
- is the training dataset
- is the entropy function
The paper discusses various ensemble methods such as deep ensembles, deep ensemble Bayesian, and UQ in traditional machine learning domain using ensemble techniques, providing an overview of their strengths and weaknesses.
The paper includes a comprehensive review of recent articles published in quantifying uncertainty in AI (ML and DL) using different approaches. The review identifies main research gaps of UQ methods and pointing out solid future directions for researchers in this domain.