Awareness of uncertainty in classification using a multivariate model and multi-views (2404.10314v1)
Abstract: One of the ways to make artificial intelligence more natural is to give it some room for doubt. Two main questions should be resolved in that way. First, how to train a model to estimate uncertainties of its own predictions? And then, what to do with the uncertain predictions if they appear? First, we proposed an uncertainty-aware negative log-likelihood loss for the case of N-dimensional multivariate normal distribution with spherical variance matrix to the solution of N-classes classification tasks. The loss is similar to the heteroscedastic regression loss. The proposed model regularizes uncertain predictions, and trains to calculate both the predictions and their uncertainty estimations. The model fits well with the label smoothing technique. Second, we expanded the limits of data augmentation at the training and test stages, and made the trained model to give multiple predictions for a given number of augmented versions of each test sample. Given the multi-view predictions together with their uncertainties and confidences, we proposed several methods to calculate final predictions, including mode values and bin counts with soft and hard weights. For the latter method, we formalized the model tuning task in the form of multimodal optimization with non-differentiable criteria of maximum accuracy, and applied particle swarm optimization to solve the tuning task. The proposed methodology was tested using CIFAR-10 dataset with clean and noisy labels and demonstrated good results in comparison with other uncertainty estimation methods related to sample selection, co-teaching, and label smoothing.
- A review of uncertainty quantification in deep learning: Techniques, applications and challenges. Information fusion, 76:243–297, 2021.
- Pitfalls of in-domain uncertainty estimation and ensembling in deep learning. arXiv preprint arXiv:2002.06470, 2020.
- A simple probabilistic method for deep classification under input-dependent label noise. arXiv preprint arXiv:2003.06778, 2020.
- A. Dai and M. Nießner. 3dmv: Joint 3d-multi-view prediction for 3d semantic scene segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), pages 452–468, 2018.
- Logistic-normal likelihoods for heteroscedastic label noise, 2023.
- Deep Learning. MIT Press, 2016. http://www.deeplearningbook.org.
- Co-teaching: Robust training of deep neural networks with extremely noisy labels, 2018.
- L. Hui and M. Belkin. Evaluation of neural architectures trained with square loss vs cross-entropy in classification tasks, 2021.
- Beyond synthetic noise: Deep learning on controlled noisy labels, 2020.
- A. Kendall and Y. Gal. What uncertainties do we need in bayesian deep learning for computer vision? Advances in neural information processing systems, 30, 2017.
- A. Krizhevsky. Learning multiple layers of features from tiny images. 2009. URL https://api.semanticscholar.org/CorpusID:18268744.
- Verified uncertainty calibration, 2020.
- Packed-ensembles for efficient uncertainty estimation, 2023.
- On interaction between augmentations and corruptions in natural corruption robustness, 2021.
- Deep neural networks are easily fooled: High confidence predictions for unrecognizable images, 2015.
- Can you trust your model’s uncertainty? evaluating predictive uncertainty under dataset shift, 2019.
- Uncertainty in neural networks: Approximately bayesian ensembling. In International conference on artificial intelligence and statistics, pages 234–244. PMLR, 2020.
- Understanding softmax confidence and uncertainty. CoRR, abs/2106.04972, 2021. URL https://arxiv.org/abs/2106.04972.
- S. Prince. Computer Vision: Models Learning and Inference. Cambridge University Press, 2012.
- S. J. Prince. Understanding Deep Learning. MIT Press, 2023. URL http://udlbook.com.
- Gaussian Processes for Machine Learning. MIT Press, 2006.
- Data augmentation can improve robustness. Advances in Neural Information Processing Systems, 34:29935–29948, 2021.
- On the pitfalls of heteroscedastic uncertainty estimation with probabilistic neural networks. arXiv preprint arXiv:2203.09168, 2022.
- Evidential deep learning to quantify classification uncertainty, 2018.
- Particle swarm optimization: A comprehensive survey. IEEE Access, 10:10031–10061, 2022.
- Y. Shi and R. C. Eberhart. Empirical study of particle swarm optimization. In Proceedings of the 1999 congress on evolutionary computation-CEC99 (Cat. No. 99TH8406), volume 3, pages 1945–1950. IEEE, 1999.
- Particle swarm optimization algorithm: an overview. Soft computing, 22:387–408, 2018.
- To smooth or not? when label smoothing meets noisy labels. arXiv preprint arXiv:2106.04149, 2021.
- Sample selection with uncertainty of losses for learning with noisy labels, 2021.
- Searching to exploit memorization effect in learning from corrupted labels, 2020.