Papers
Topics
Authors
Recent
Search
2000 character limit reached

On the Road to Clarity: Exploring Explainable AI for World Models in a Driver Assistance System

Published 26 Apr 2024 in cs.LG, cs.CV, and cs.MA | (2404.17350v1)

Abstract: In Autonomous Driving (AD) transparency and safety are paramount, as mistakes are costly. However, neural networks used in AD systems are generally considered black boxes. As a countermeasure, we have methods of explainable AI (XAI), such as feature relevance estimation and dimensionality reduction. Coarse graining techniques can also help reduce dimensionality and find interpretable global patterns. A specific coarse graining method is Renormalization Groups from statistical physics. It has previously been applied to Restricted Boltzmann Machines (RBMs) to interpret unsupervised learning. We refine this technique by building a transparent backbone model for convolutional variational autoencoders (VAE) that allows mapping latent values to input features and has performance comparable to trained black box VAEs. Moreover, we propose a custom feature map visualization technique to analyze the internal convolutional layers in the VAE to explain internal causes of poor reconstruction that may lead to dangerous traffic scenarios in AD applications. In a second key contribution, we propose explanation and evaluation techniques for the internal dynamics and feature relevance of prediction networks. We test a long short-term memory (LSTM) network in the computer vision domain to evaluate the predictability and in future applications potentially safety of prediction models. We showcase our methods by analyzing a VAE-LSTM world model that predicts pedestrian perception in an urban traffic situation.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (34)
  1. A. Erasmus, T. D. P. Brunet, and E. Fisher, “What is interpretability?” Philosophy & Technology, vol. 34, no. 4, pp. 833–862, Dec 2021.
  2. G. Yannis, D. Nikolaou, A. Laiou, Y. A. Stürmer, I. Buttler, and D. Jankowska-Karpa, “Vulnerable road users: Cross-cultural perspectives on performance and attitudes,” IATSS research, vol. 44, no. 3, pp. 220–229, 2020.
  3. L. Lévêque, M. Ranchet, J. Deniel, J.-C. Bornard, and T. Bellet, “Where do pedestrians look when crossing? a state of the art of the eye-tracking studies,” IEEE Access, vol. 8, pp. 164 833–164 843, 2020.
  4. J. Petzold, M. Wahby, F. Stark, U. Behrje, and H. Hamann, ““if you could see me through my eyes”: Predicting pedestrian perception,” in 8th Int. Conf. on Control, Automation & Robotics (ICCAR), 2022, pp. 184–190.
  5. A. Dosovitskiy, G. Ros, F. Codevilla, A. Lopez, and V. Koltun, “CARLA: An open urban driving simulator,” in Proceedings of the 1st Annual Conference on Robot Learning, 2017, pp. 1–16.
  6. Y. Li, D. Ma, Z. An, Z. Wang, Y. Zhong, S. Chen, and C. Feng, “V2X-Sim: Multi-agent collaborative perception dataset and benchmark for autonomous driving,” IEEE Robotics and Automation Letters, vol. 7, no. 4, pp. 10 914–10 921, 2022.
  7. D. P. Kingma and M. Welling, “Auto-Encoding Variational Bayes,” in 2nd Int. Conf.e on Learning Representations, ICLR, 2014.
  8. D. Ha and J. Schmidhuber, “Recurrent world models facilitate policy evolution,” in Proceedings of the 32nd International Conference on Neural Information Processing Systems (NIPS 2018).   USA: Curran Associates Inc., 2018, pp. 2455–2467.
  9. A. De Mello Koch, E. De Mello Koch, and R. De Mello Koch, “Why unsupervised deep networks generalize,” vol. abs/2012.03531, 2020. [Online]. Available: https://arxiv.org/abs/2012.03531
  10. G. Montavon, W. Samek, and K.-R. Müller, “Methods for interpreting and understanding deep neural networks,” Digital Signal Processing, vol. 73, pp. 1–15, 2018.
  11. A. Barredo Arrieta, N. Díaz-Rodríguez, J. Del Ser, A. Bennetot, S. Tabik, A. Barbado, S. Garcia, S. Gil-Lopez, D. Molina, R. Benjamins, R. Chatila, and F. Herrera, “Explainable artificial intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible ai,” Information Fusion, vol. 58, pp. 82–115, 2020.
  12. S. Atakishiyev, M. Salameh, H. Yao, and R. Goebel, “Towards safe, explainable, and regulated autonomous driving,” arXiv preprint arXiv:2111.10518, 2021.
  13. D. Omeiza, H. Webb, M. Jirotka, and L. Kunze, “Explanations in autonomous driving: A survey,” IEEE Transactions on Intelligent Transportation Systems, 2021.
  14. É. Zablocki, H. Ben-Younes, P. Pérez, and M. Cord, “Explainability of deep vision-based autonomous driving systems: Review and challenges,” International Journal of Computer Vision, pp. 1–28, 2022.
  15. J. Dong, S. Chen, S. Zong, T. Chen, and S. Labi, “Image transformer for explainable autonomous driving system,” in 2021 IEEE International Intelligent Transportation Systems Conference (ITSC).   IEEE, 2021, pp. 2732–2737.
  16. N. T. S. Board, “Collision between vehicle controlled by developmental automated driving system and pedestrian in tempe, arizona, on march 18, 2018,” 2019.
  17. I. Rahwan, M. Cebrian, N. Obradovich et al., “Machine behaviour,” Nature, vol. 568, p. 477–486, 2019.
  18. A. Karpathy, J. Johnson, and L. Fei-Fei, “Visualizing and understanding recurrent networks,” 2015. [Online]. Available: https://arxiv.org/abs/1506.02078
  19. S. Bach, A. Binder, G. Montavon, F. Klauschen, K.-R. Müller, and W. Samek, “On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation,” PLOS ONE, vol. 10, no. 7, pp. 1–46, 07 2015.
  20. C. Olah, A. Mordvintsev, and L. Schubert, “Feature visualization,” Distill, 2017, https://distill.pub/2017/feature-visualization.
  21. R. R. Selvaraju, M. Cogswell, A. Das, R. Vedantam, D. Parikh, and D. Batra, “Grad-cam: Visual explanations from deep networks via gradient-based localization,” in 2017 IEEE International Conference on Computer Vision (ICCV), 2017, pp. 618–626.
  22. W. Liu, R. Li, M. Zheng, S. Karanam, Z. Wu, B. Bhanu, R. J. Radke, and O. Camps, “Towards visually explaining variational autoencoders,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 8642–8651.
  23. M. B. Muhammad and M. Yeasin, “Eigen-cam: Class activation map using principal components,” in 2020 International Joint Conference on Neural Networks (IJCNN).   IEEE, Jul. 2020.
  24. I. Cetin, M. Stephens, O. Camara, and M. A. G. Ballester, “Attri-VAE: Attribute-based interpretable representations of medical images with variational autoencoders,” Computerized Medical Imaging and Graphics, vol. 104, p. 102158, 2023.
  25. I. Higgins, L. Matthey, A. Pal, C. Burgess, X. Glorot, M. Botvinick, S. Mohamed, and A. Lerchner, “beta-vae: Learning basic visual concepts with a constrained variational framework,” in International conference on learning representations, 2017.
  26. S. Iso, S. Shiba, and S. Yokoo, “Scale-invariant feature extraction of neural network and renormalization group flow,” Phys. Rev. E, vol. 97, p. 053304, May 2018.
  27. P. Mehta and D. J. Schwab, “An exact mapping between the variational renormalization group and deep learning,” ArXiv, vol. abs/1410.3831, 2014.
  28. E. De Mello Koch, R. De Mello Koch, and L. Cheng, “Is deep learning a renormalization group flow?” IEEE Access, vol. 8, pp. 106 487–106 505, 2020.
  29. S. Hochreiter and J. Schmidhuber, “Long short-term memory,” Neural Comput., vol. 9, no. 8, p. 1735–1780, nov 1997.
  30. M. M. Ahsan, M. A. P. Mahmud, P. K. Saha, K. D. Gupta, and Z. Siddique, “Effect of data scaling methods on machine learning algorithms and model performance,” Technologies, 2021.
  31. C. Rudin, “Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead,” Nature Machine Intelligence, vol. 1, no. 5, pp. 206–215, May 2019.
  32. S. Kullback and R. A. Leibler, “On information and sufficiency,” The annals of mathematical statistics, vol. 22, no. 1, pp. 79–86, 1951.
  33. L. Arras, G. Montavon, K.-R. Müller, and W. Samek, “Explaining recurrent neural network predictions in sentiment analysis,” in Proceedings of the 8th Workshop on Computational Approaches to Subjectivity, Sentiment and Social Media Analysis, 2017, pp. 159–168.
  34. A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. u. Kaiser, and I. Polosukhin, “Attention is all you need,” in Advances in Neural Information Processing Systems, I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, Eds., vol. 30.   Curran Associates, Inc., 2017.

Summary

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 2 tweets with 6 likes about this paper.