A Trustworthiness Score to Evaluate DNN Predictions (2301.08839v6)
Abstract: Due to the black box nature of deep neural networks (DNN), the continuous validation of DNN during operation is challenging with the absence of a human monitor. As a result this makes it difficult for developers and regulators to gain confidence in the deployment of autonomous systems employing DNN. It is critical for safety during operation to know when DNN's predictions are trustworthy or suspicious. With the absence of a human monitor, the basic approach is to use the model's output confidence score to assess if predictions are trustworthy or suspicious. However, the model's confidence score is a result of computations coming from a black box, therefore lacks transparency and makes it challenging to automatedly credit trustworthiness to predictions. We introduce the trustworthiness score (TS), a simple metric that provides a more transparent and effective way of providing confidence in DNN predictions compared to model's confidence score. The metric quantifies the trustworthiness in a prediction by checking for the existence of certain features in the predictions made by the DNN. We also use the underlying idea of the TS metric, to provide a suspiciousness score (SS) in the overall input frame to help in the detection of suspicious frames where false negatives exist. We conduct a case study using YOLOv5 on persons detection to demonstrate our method and usage of TS and SS. The case study shows that using our method consistently improves the precision of predictions compared to relying on model confidence score alone, for both 1) approving of trustworthy predictions (~20% improvement) and 2) detecting suspicious frames (~5% improvement).
- “A review of uncertainty quantification in deep learning: Techniques, applications and challenges” In Information Fusion 76, 2021
- “On Specifying for Trustworthiness”, 2022 arXiv:2206.11421
- “Getting a CLUE: A Method for Explaining Uncertainty Estimates”, 2020 arXiv:2006.06848
- “The oracle problem in software testing: A survey” In IEEE Transactions on Software Engineering 41.5 IEEE, 2015 DOI: 10.1109/TSE.2014.2372785
- Paul Bauer “Conceptualizing Trust and Trustworthiness” In Ssrn, 2013 DOI: 10.2139/ssrn.2325989
- Kim Bie, Ana Lucic and Hinda Haned “To Trust or Not to Trust a Regressor: Estimating and Explaining Trustworthiness of Regression Predictions”, 2021 arXiv:2104.06982
- “Industry Internet of Things Vocabulary” In Industry IoT Consortium, 2022
- “Histograms of oriented gradients for human detection” In 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05) 1, 2005 DOI: 10.1109/CVPR.2005.177
- “Towards A Rigorous Science of Interpretable Machine Learning”, 2017
- Chin-Shyurng Fahn, Chu-Ping Lee and Yu-Shu Yeh “A real-time pedestrian legs detection and tracking system used for autonomous mobile robots” In 2017 International Conference on Applied System Innovation (ICASI), 2017 DOI: 10.1109/ICASI.2017.7988208
- Ruth Fong, Mandela Patrick and Andrea Vedaldi “Understanding deep networks via extremal perturbations and smooth masks” In Proceedings of the IEEE International Conference on Computer Vision, 2019 DOI: 10.1109/ICCV.2019.00304
- “Dropout as a bayesian approximation: Representing model uncertainty in deep learning” In international conference on machine learning, 2016 PMLR
- Zoubin Ghahramani “Probabilistic machine learning and artificial intelligence” In Nature 521.7553 Nature Publishing Group, 2015
- “Operational Adaptation of DNN Classifiers using Elastic Weight Consolidation”, 2022 DOI: 10.48550/ARXIV.2205.00147
- “Safety Validation of Autonomous Vehicles using Assertion Checking” arXiv, 2021 arXiv:2111.04611
- “Guidance on the assurance of machine learning in autonomous systems (AMLAS)”, 2021 arXiv:2102.01564
- “Machine learning with a reject option: A survey”, 2021 arXiv:2107.11277
- Bernhard Hommel and Lorenza S. Colzato “Interpersonal trust: an event-based account” In Frontiers in Psychology, 2015 DOI: 10.3389/fpsyg.2015.01399
- Darryl Hond, Hamid Asgari and Daniel Jeffery “Verifying Artificial Neural Network Classifier Performance Using Dataset Dissimilarity Measures” In Proceedings - 19th ICMLA IEEE, 2020 DOI: 10.1109/ICMLA51294.2020.00027
- “To trust or not to trust a classifier” In Advances in Neural Information Processing Systems, 2018 arXiv:1805.11783
- Glenn Jocher “YOLOv5 by Ultralytics”, 2020 DOI: 10.5281/zenodo.3908559
- “One millisecond face alignment with an ensemble of regression trees” In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition IEEE, 2014 DOI: 10.1109/CVPR.2014.241
- Davis E King “Max-margin object detection” In arXiv preprint arXiv:1502.00046, 2015
- “Uncertainty Estimation in One-Stage Object Detection” In 2019 IEEE Intelligent Transportation Systems Conference, ITSC 2019, 2019 DOI: 10.1109/ITSC.2019.8917494
- “Microsoft COCO: Common Objects in Context” In CoRR abs/1405.0312, 2014
- Zachary C. Lipton “The mythos of model interpretability” In Communications of the ACM 61.10, 2018 DOI: 10.1145/3233231
- “Human leg detection from depth sensing” In 2018 IEEE International Conference on Automation, Quality and Testing, Robotics, AQTR 2018 - THETA 21st Edition, Proceedings IEEE, 2018 DOI: 10.1109/AQTR.2018.8402735
- “MediaPipe: A Framework for Building Perception Pipelines” arXiv, 2019 DOI: 10.48550/ARXIV.1906.08172
- Scott M. Lundberg and Su In Lee “A unified approach to interpreting model predictions” In Advances in Neural Information Processing Systems, 2017 arXiv:1705.07874
- Panagiotis Mitkidis, Michaela Porubanova and Andreas Roepstorff “Editorial: Trust: The Limits of Human Moral” In Frontiers in Psychology 8.3, 2017 DOI: 10.3389/fpsyg.2017.00178
- Warsun Najib, Selo Sulistyo and Widyawan “Survey on trust calculation methods in internet of things” In Procedia Computer Science Elsevier B.V., 2019 DOI: 10.1016/j.procs.2019.11.245
- “Predicting YOLO Misdetection by Learning Grid Cell Consensus” In Proceedings - 20th ICMLA IEEE, 2021 DOI: 10.1109/ICMLA52953.2021.00107
- “You Only Look Once: Unified, Real-Time Object Detection” In CoRR abs/1506.02640, 2015 arXiv:1506.02640
- Marco Tulio Ribeiro, Sameer Singh and Carlos Guestrin “”Why Should I Trust You?” Explaining the Predictions of Any Classifier” In NAACL-HLT 2016, 2016 DOI: 10.18653/v1/n16-3020
- “Interpretable machine learning: Fundamental principles and 10 grand challenges” In Statistics Surveys 16, 2022 DOI: 10.1214/21-SS133
- “Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization” In International Journal of Computer Vision 128.2, 2020 DOI: 10.1007/s11263-019-01228-7
- “Explaining Image Classifiers Using Statistical Fault Localization” In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) 12373 LNCS, 2020 DOI: 10.1007/978-3-030-58604-1˙24
- “Systems and software engineering — Vocabulary ISO/IEC/IEEE 24765:2017(E)”, 2017
- “Systems Engineering Handbook - NASA/SP-2016-6105”, 2016
- Scott Thiebes, Sebastian Lins and Ali Sunyaev “Trustworthy artificial intelligence” In Electronic Markets 31 Springer, 2021, pp. 447–464
- “Plex: Towards reliability using pretrained large model extensions”, 2022 arXiv:2207.07411
- “Trustworthy AI Development Guidelines for Human System Interaction” In 2020 13th International Conference on Human System Interaction (HSI), 2020, pp. 130–136 DOI: 10.1109/HSI49210.2020.9142644
- Jeannette M. Wing “Trustworthy AI” In Communications of the ACM 64.10, 2021 DOI: 10.1145/3448248