Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
167 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Error-Driven Uncertainty Aware Training (2405.01205v2)

Published 2 May 2024 in cs.LG and cs.CV

Abstract: Neural networks are often overconfident about their predictions, which undermines their reliability and trustworthiness. In this work, we present a novel technique, named Error-Driven Uncertainty Aware Training (EUAT), which aims to enhance the ability of neural classifiers to estimate their uncertainty correctly, namely to be highly uncertain when they output inaccurate predictions and low uncertain when their output is accurate. The EUAT approach operates during the model's training phase by selectively employing two loss functions depending on whether the training examples are correctly or incorrectly predicted by the model. This allows for pursuing the twofold goal of i) minimizing model uncertainty for correctly predicted inputs and ii) maximizing uncertainty for mispredicted inputs, while preserving the model's misprediction rate. We evaluate EUAT using diverse neural models and datasets in the image recognition domains considering both non-adversarial and adversarial settings. The results show that EUAT outperforms existing approaches for uncertainty estimation (including other uncertainty-aware training techniques, calibration, ensembles, and DEUP) by providing uncertainty estimates that not only have higher quality when evaluated via statistical metrics (e.g., correlation with residuals) but also when employed to build binary classifiers that decide whether the model's output can be trusted or not and under distributional data shifts.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (46)
  1. A review of uncertainty quantification in deep learning: Techniques, applications and challenges. Information Fusion, 76:243–297, 2021.
  2. Depth uncertainty in neural networks. In Proceedings of the 34th International Conference on Neural Information Processing Systems, NIPS’20, Red Hook, NY, USA, 2020. Curran Associates Inc.
  3. Objective evaluation of deep uncertainty predictions for covid-19 detection. Scientific Reports, 12, 2022.
  4. Single shot mc dropout approximation. ArXiv, abs/2007.03293, 2020. URL https://api.semanticscholar.org/CorpusID:220381176.
  5. N. Carlini and D. Wagner. Adversarial examples are not easily detected: Bypassing ten detection methods. In Proceedings of the 10th ACM Workshop on Artificial Intelligence and Security, AISec ’17, page 3–14, New York, NY, USA, 2017. Association for Computing Machinery. ISBN 9781450352024. URL https://doi.org/10.1145/3128572.3140444.
  6. A. Damianou and N. D. Lawrence. Deep Gaussian processes. In Proceedings of the Sixteenth International Conference on Artificial Intelligence and Statistics, volume 31 of Proceedings of Machine Learning Research, pages 207–215, Scottsdale, Arizona, USA, 2013. PMLR.
  7. Uncertainty-aware training of neural networks for selective medical image segmentation. In Proceedings of the Third Conference on Medical Imaging with Deep Learning, volume 121 of Proceedings of Machine Learning Research, pages 156–173. PMLR, 2020.
  8. Training uncertainty-aware classifiers with conformalized deep learning. In S. Koyejo, S. Mohamed, A. Agarwal, D. Belgrave, K. Cho, and A. Oh, editors, Advances in Neural Information Processing Systems, volume 35, 2022.
  9. Detecting adversarial samples from artifacts. In International Conference on Machine Learning, 2017.
  10. Y. Gal and Z. Ghahramani. Dropout as a bayesian approximation: Representing model uncertainty in deep learning. In Proceedings of The 33rd International Conference on Machine Learning, volume 48. PMLR, 2016.
  11. A survey of uncertainty in deep neural networks. Artificial Intelligence Review, 56, 2023.
  12. S. J. Godsill. On the relationship between markov chain monte carlo methods for model uncertainty. Journal of Computational and Graphical Statistics, 10(2):230–248, 2001.
  13. Explaining and harnessing adversarial examples. In ICLR, 2015.
  14. On calibration of modern neural networks. In Proceedings of the 34th International Conference on Machine Learning, volume 70. PMLR, 2017.
  15. Calibration of neural networks using splines. In International Conference on Learning Representations, 2021.
  16. Deep residual learning for image recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016.
  17. D. Hendrycks and T. Dietterich. Benchmarking neural network robustness to common corruptions and perturbations. In International Conference on Learning Representations, 2019.
  18. Soft calibration objectives for neural networks. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P. Liang, and J. W. Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, 2021.
  19. Cua loss: Class uncertainty-aware gradient modulation for robust object detection. IEEE Transactions on Circuits and Systems for Video Technology, 31(9), 2021.
  20. R. Krishnan and O. Tickoo. Improving model calibration with accuracy versus uncertainty optimization. In Advances in Neural Information Processing Systems, volume 33, 2020.
  21. A. Krizhevsky and G. Hinton. Learning multiple layers of features from tiny images. Technical report, University of Toronto, 2009.
  22. Beta calibration: a well-founded and easily implemented improvement on logistic calibration for binary classifiers. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, volume 54, 2017.
  23. DEUP: Direct epistemic uncertainty prediction. Transactions on Machine Learning Research, 2023. ISSN 2835-8856. URL https://openreview.net/forum?id=eGLdVRvvfQ.
  24. Simple and scalable predictive uncertainty estimation using deep ensembles. In Proceedings of the 31st International Conference on Neural Information Processing Systems, 2017.
  25. Leveraging uncertainty information from deep neural networks for disease detection. Scientific reports, 7(1), 2017.
  26. The devil is in the margin: Margin-based label smoothing for network calibration. In Computer Vision and Pattern Recognition Conference, 2022.
  27. Class adaptive network calibration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 16070–16079, June 2023.
  28. D. J. C. MacKay. A Practical Bayesian Framework for Backpropagation Networks. Neural Computation, 4(3):448–472, 1992.
  29. D. J. C. MacKay. Information Theory, Inference, and Learning Algorithms. Cambridge University Press, 2003.
  30. Towards deep learning models resistant to adversarial attacks. In ICLR, 2018.
  31. TCE: A test-based approach to measuring calibration error. In Proceedings of the Thirty-Ninth Conference on Uncertainty in Artificial Intelligence, 2023.
  32. Calibrating deep neural networks using focal loss. In Advances in Neural Information Processing Systems, volume 33, 2020.
  33. Obtaining well calibrated probabilities using bayesian binning. In Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, 2015.
  34. R. M. Neal. Bayesian Learning for Neural Networks. Springer-Verlag, 1996.
  35. Reading digits in natural images with unsupervised feature learning. In NIPS Workshop on Deep Learning and Unsupervised Feature Learning 2011, 2011.
  36. Measuring calibration in deep learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, June 2019.
  37. Can you trust your model's uncertainty? evaluating predictive uncertainty under dataset shift. In Advances in Neural Information Processing Systems, volume 32, 2019.
  38. Uncertainty in neural networks: Approximately bayesian ensembling. In Proceedings of the Twenty Third International Conference on Artificial Intelligence and Statistics, 2020.
  39. J. C. Platt. Probabilistic outputs for support vector machines and comparisons to regularized likelihood methods. In Advances in Large Margin Classifiers, 1999.
  40. ImageNet Large Scale Visual Recognition Challenge. International Journal of Computer Vision (IJCV), 115(3):211–252, 2015. 10.1007/s11263-015-0816-y.
  41. An uncertainty-aware loss function for training neural networks with calibrated predictions. ArXiv, abs/2110.03260, 2023.
  42. L. Smith and Y. Gal. Understanding measures of uncertainty for adversarial example detection. In Proceedings of the Thirty-Fourth Conference on Uncertainty in Artificial Intelligence, 2018.
  43. H. Wang and D.-Y. Yeung. A survey on bayesian deep learning. ACM Comput. Surv., 53(5), sep 2020.
  44. Adversarial distillation of Bayesian neural network posteriors. In Proceedings of the 35th International Conference on Machine Learning, volume 80, 2018.
  45. B. Zadrozny and C. Elkan. Obtaining calibrated probability estimates from decision trees and naive bayesian classifiers. In Proceedings of the Eighteenth International Conference on Machine Learning, ICML ’01, 2001.
  46. S. Zagoruyko and N. Komodakis. Wide residual networks. In Proceedings of the British Machine Vision Conference 2016, 2016.

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com