Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
156 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Trustworthiness Score to Evaluate DNN Predictions (2301.08839v6)

Published 21 Jan 2023 in cs.LG

Abstract: Due to the black box nature of deep neural networks (DNN), the continuous validation of DNN during operation is challenging with the absence of a human monitor. As a result this makes it difficult for developers and regulators to gain confidence in the deployment of autonomous systems employing DNN. It is critical for safety during operation to know when DNN's predictions are trustworthy or suspicious. With the absence of a human monitor, the basic approach is to use the model's output confidence score to assess if predictions are trustworthy or suspicious. However, the model's confidence score is a result of computations coming from a black box, therefore lacks transparency and makes it challenging to automatedly credit trustworthiness to predictions. We introduce the trustworthiness score (TS), a simple metric that provides a more transparent and effective way of providing confidence in DNN predictions compared to model's confidence score. The metric quantifies the trustworthiness in a prediction by checking for the existence of certain features in the predictions made by the DNN. We also use the underlying idea of the TS metric, to provide a suspiciousness score (SS) in the overall input frame to help in the detection of suspicious frames where false negatives exist. We conduct a case study using YOLOv5 on persons detection to demonstrate our method and usage of TS and SS. The case study shows that using our method consistently improves the precision of predictions compared to relying on model confidence score alone, for both 1) approving of trustworthy predictions (~20% improvement) and 2) detecting suspicious frames (~5% improvement).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (43)
  1. “A review of uncertainty quantification in deep learning: Techniques, applications and challenges” In Information Fusion 76, 2021
  2. “On Specifying for Trustworthiness”, 2022 arXiv:2206.11421
  3. “Getting a CLUE: A Method for Explaining Uncertainty Estimates”, 2020 arXiv:2006.06848
  4. “The oracle problem in software testing: A survey” In IEEE Transactions on Software Engineering 41.5 IEEE, 2015 DOI: 10.1109/TSE.2014.2372785
  5. Paul Bauer “Conceptualizing Trust and Trustworthiness” In Ssrn, 2013 DOI: 10.2139/ssrn.2325989
  6. Kim Bie, Ana Lucic and Hinda Haned “To Trust or Not to Trust a Regressor: Estimating and Explaining Trustworthiness of Regression Predictions”, 2021 arXiv:2104.06982
  7. “Industry Internet of Things Vocabulary” In Industry IoT Consortium, 2022
  8. “Histograms of oriented gradients for human detection” In 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05) 1, 2005 DOI: 10.1109/CVPR.2005.177
  9. “Towards A Rigorous Science of Interpretable Machine Learning”, 2017
  10. Chin-Shyurng Fahn, Chu-Ping Lee and Yu-Shu Yeh “A real-time pedestrian legs detection and tracking system used for autonomous mobile robots” In 2017 International Conference on Applied System Innovation (ICASI), 2017 DOI: 10.1109/ICASI.2017.7988208
  11. Ruth Fong, Mandela Patrick and Andrea Vedaldi “Understanding deep networks via extremal perturbations and smooth masks” In Proceedings of the IEEE International Conference on Computer Vision, 2019 DOI: 10.1109/ICCV.2019.00304
  12. “Dropout as a bayesian approximation: Representing model uncertainty in deep learning” In international conference on machine learning, 2016 PMLR
  13. Zoubin Ghahramani “Probabilistic machine learning and artificial intelligence” In Nature 521.7553 Nature Publishing Group, 2015
  14. “Operational Adaptation of DNN Classifiers using Elastic Weight Consolidation”, 2022 DOI: 10.48550/ARXIV.2205.00147
  15. “Safety Validation of Autonomous Vehicles using Assertion Checking” arXiv, 2021 arXiv:2111.04611
  16. “Guidance on the assurance of machine learning in autonomous systems (AMLAS)”, 2021 arXiv:2102.01564
  17. “Machine learning with a reject option: A survey”, 2021 arXiv:2107.11277
  18. Bernhard Hommel and Lorenza S. Colzato “Interpersonal trust: an event-based account” In Frontiers in Psychology, 2015 DOI: 10.3389/fpsyg.2015.01399
  19. Darryl Hond, Hamid Asgari and Daniel Jeffery “Verifying Artificial Neural Network Classifier Performance Using Dataset Dissimilarity Measures” In Proceedings - 19th ICMLA IEEE, 2020 DOI: 10.1109/ICMLA51294.2020.00027
  20. “To trust or not to trust a classifier” In Advances in Neural Information Processing Systems, 2018 arXiv:1805.11783
  21. Glenn Jocher “YOLOv5 by Ultralytics”, 2020 DOI: 10.5281/zenodo.3908559
  22. “One millisecond face alignment with an ensemble of regression trees” In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition IEEE, 2014 DOI: 10.1109/CVPR.2014.241
  23. Davis E King “Max-margin object detection” In arXiv preprint arXiv:1502.00046, 2015
  24. “Uncertainty Estimation in One-Stage Object Detection” In 2019 IEEE Intelligent Transportation Systems Conference, ITSC 2019, 2019 DOI: 10.1109/ITSC.2019.8917494
  25. “Microsoft COCO: Common Objects in Context” In CoRR abs/1405.0312, 2014
  26. Zachary C. Lipton “The mythos of model interpretability” In Communications of the ACM 61.10, 2018 DOI: 10.1145/3233231
  27. “Human leg detection from depth sensing” In 2018 IEEE International Conference on Automation, Quality and Testing, Robotics, AQTR 2018 - THETA 21st Edition, Proceedings IEEE, 2018 DOI: 10.1109/AQTR.2018.8402735
  28. “MediaPipe: A Framework for Building Perception Pipelines” arXiv, 2019 DOI: 10.48550/ARXIV.1906.08172
  29. Scott M. Lundberg and Su In Lee “A unified approach to interpreting model predictions” In Advances in Neural Information Processing Systems, 2017 arXiv:1705.07874
  30. Panagiotis Mitkidis, Michaela Porubanova and Andreas Roepstorff “Editorial: Trust: The Limits of Human Moral” In Frontiers in Psychology 8.3, 2017 DOI: 10.3389/fpsyg.2017.00178
  31. Warsun Najib, Selo Sulistyo and Widyawan “Survey on trust calculation methods in internet of things” In Procedia Computer Science Elsevier B.V., 2019 DOI: 10.1016/j.procs.2019.11.245
  32. “Predicting YOLO Misdetection by Learning Grid Cell Consensus” In Proceedings - 20th ICMLA IEEE, 2021 DOI: 10.1109/ICMLA52953.2021.00107
  33. “You Only Look Once: Unified, Real-Time Object Detection” In CoRR abs/1506.02640, 2015 arXiv:1506.02640
  34. Marco Tulio Ribeiro, Sameer Singh and Carlos Guestrin “”Why Should I Trust You?” Explaining the Predictions of Any Classifier” In NAACL-HLT 2016, 2016 DOI: 10.18653/v1/n16-3020
  35. “Interpretable machine learning: Fundamental principles and 10 grand challenges” In Statistics Surveys 16, 2022 DOI: 10.1214/21-SS133
  36. “Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization” In International Journal of Computer Vision 128.2, 2020 DOI: 10.1007/s11263-019-01228-7
  37. “Explaining Image Classifiers Using Statistical Fault Localization” In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) 12373 LNCS, 2020 DOI: 10.1007/978-3-030-58604-1˙24
  38. “Systems and software engineering — Vocabulary ISO/IEC/IEEE 24765:2017(E)”, 2017
  39. “Systems Engineering Handbook - NASA/SP-2016-6105”, 2016
  40. Scott Thiebes, Sebastian Lins and Ali Sunyaev “Trustworthy artificial intelligence” In Electronic Markets 31 Springer, 2021, pp. 447–464
  41. “Plex: Towards reliability using pretrained large model extensions”, 2022 arXiv:2207.07411
  42. “Trustworthy AI Development Guidelines for Human System Interaction” In 2020 13th International Conference on Human System Interaction (HSI), 2020, pp. 130–136 DOI: 10.1109/HSI49210.2020.9142644
  43. Jeannette M. Wing “Trustworthy AI” In Communications of the ACM 64.10, 2021 DOI: 10.1145/3448248
Citations (2)

Summary

We haven't generated a summary for this paper yet.