ReMAV: Reward Modeling of Autonomous Vehicles for Finding Likely Failure Events (2308.14550v2)
Abstract: Autonomous vehicles are advanced driving systems that are well known to be vulnerable to various adversarial attacks, compromising vehicle safety and posing a risk to other road users. Rather than actively training complex adversaries by interacting with the environment, there is a need to first intelligently find and reduce the search space to only those states where autonomous vehicles are found to be less confident. In this paper, we propose a black-box testing framework ReMAV that uses offline trajectories first to analyze the existing behavior of autonomous vehicles and determine appropriate thresholds to find the probability of failure events. To this end, we introduce a three-step methodology which i) uses offline state action pairs of any autonomous vehicle under test, ii) builds an abstract behavior representation using our designed reward modeling technique to analyze states with uncertain driving decisions, and iii) uses a disturbance model for minimal perturbation attacks where the driving decisions are less confident. Our reward modeling technique helps in creating a behavior representation that allows us to highlight regions of likely uncertain behavior even when the standard autonomous vehicle performs well. We perform our experiments in a high-fidelity urban driving environment using three different driving scenarios containing single- and multi-agent interactions. Our experiment shows an increase in 35, 23, 48, and 50% in the occurrences of vehicle collision, road object collision, pedestrian collision, and offroad steering events, respectively by the autonomous vehicle under test, demonstrating a significant increase in failure events. We compare ReMAV with two baselines and show that ReMAV demonstrates significantly better effectiveness in generating failure events compared to the baselines in all evaluation metrics.
- J. Garcia, Y. Feng, J. Shen, S. Almanee, Y. Xia, Chen, and Q. Alfred, “A comprehensive study of autonomous vehicle bugs,” in Proceedings of the ACM/IEEE 42nd International Conference on Software Engineering, ser. ICSE ’20, 2020.
- V. Riccio, G. Jahangirova, and A. e. a. Stocco, “Testing machine learning based systems: a systematic mapping,” Empir Software Eng, p. 5193, 2020.
- D. Marijan and A. Gotlieb, “Software testing for machine learning,” Proceedings of the AAAI Conference on Artificial Intelligence, 2020.
- K. L. Tan, Y. Esfandiari, X. Y. Lee, Aakanksha, and S. Sarkar, “Robustifying reinforcement learning agents via action space adversarial training,” in ACC, 2020.
- X. Y. Lee, Y. Esfandiari, K. L. Tan, and S. Sarkar, “Query-based targeted action-space adversarial policies on deep reinforcement learning agents,” in ACM/IEEE 12th International Conference on Cyber-Physical Systems (ICCPS), 2021.
- B. Chalaki, L. E. Beaver, B. Remer, K. Jang, E. Vinitsky, A. M. Bayen, and A. A. Malikopoulos, “Zero-shot autonomous vehicle policy transfer: From simulation to real-world via adversarial learning,” in ICCA, 2020.
- A. Wachi, “Failure-scenario maker for rule-based agent using multi-agent adversarial reinforcement learning and its application to autonomous driving,” in IJCAI, 2019.
- Y. Abeysirigoonawardena, F. Shkurti, and G. Dudek, “Generating adversarial driving scenarios in high-fidelity simulators,” in ICRA, 2019.
- J. Yang, A. Boloor, A. Chakrabarti, X. Zhang, and Y. Vorobeychik, “Finding physical adversarial examples for autonomous driving with fast and differentiable image compositing,” in arXiv, 2020.
- M. Uřičář, P. Křížek, D. Hurych, I. Sobh, S. Yogamani, and P. Denny, “Yes, we gan: Applying adversarial techniques for autonomous driving,” in Society for Imaging Science and Technology, 2019.
- K. Jang, E. Vinitsky, B. Chalaki, B. Remer, L. Beaver, A. A. Malikopoulos, and A. Bayen, “Simulation to scaled city: Zero-shot policy transfer for traffic control via autonomous vehicles,” in 10th ACM/IEEE ICCPS, 2019.
- F. Codevilla, A. M. Lopez, V. Koltun, and A. Dosovitskiy, “On offline evaluation of vision-based driving models,” in Proceedings of the European Conference on Computer Vision (ECCV), 2018, pp. 236–251.
- F. Codevilla, E. Santana, A. M. López, and A. Gaidon, “Exploring the limitations of behavior cloning for autonomous driving,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019.
- A. Gleave, M. Dennis, C. Wild, N. Kant, S. Levine, and S. Russell, “Adversarial policies: Attacking deep reinforcement learning,” in ICLR, 2020.
- A. Corso, P. Du, K. Driggs-Campbell, and M. J. Kochenderfer, “Adaptive stress testing with reward augmentation for autonomous vehicle validation,” in ITSC, 2019.
- M. Koren, A. Nassar, and M. J. Kochenderfer, “Finding failures in high-fidelity simulation using adaptive stress testing and the backward algorithm,” in 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2021.
- M. Koren, S. Alsaif, R. Lee, and M. J. Kochenderfer, “Adaptive stress testing for autonomous vehicles,” in IEEE Intelligent Vehicles Symposium (IV), 2018.
- A. Sharif and D. Marijan, “Adversarial deep reinforcement learning for improving the robustness of multi-agent autonomous driving policies,” in 2022 29th Asia-Pacific Software Engineering Conference (APSEC), 2022.
- Y.-C. Lin, Z.-W. Hong, Y.-H. Liao, M.-L. Shih, M.-Y. Liu, and M. Sun, “Tactics of adversarial attack on deep reinforcement learning agents,” arXiv preprint arXiv:1703.06748, 2017.
- J. Schulman, F. Wolski, P. Dhariwal, A. Radford, and O. Klimov, “Proximal policy optimization algorithms,” CoRR, 2017.
- J. Fu, K. Luo, and S. Levine, “Learning robust rewards with adversarial inverse reinforcement learning,” arXiv preprint arXiv:1710.11248, 2017.
- K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” arXiv preprint arXiv:1409.1556, 2014.
- “3rd maps and navigation,” 2021. [Online]. Available: https://carla.readthedocs.io/en/latest/core_map/
- “Attacking vision-based perception in end-to-end autonomous driving models,” in Journal of Systems Architecture, 2020.
- D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” in ICLR, Y. Bengio and Y. LeCun, Eds., 2015.
- P. Moritz, R. Nishihara, S. Wang, A. Tumanov, R. Liaw, E. Liang, W. Paul, M. I. Jordan, and I. Stoica, “Ray: A distributed framework for emerging AI applications,” CoRR, 2017.
- A. Dosovitskiy, G. Ros, F. Codevilla, A. Lopez, and V. Koltun, “CARLA: An open urban driving simulator,” in CoRL, 2017.
- A. Sharif and D. Marijan, “Evaluating the robustness of deep reinforcement learning for autonomous policies in a multi-agent urban driving environment,” in 2022 IEEE 22nd International Conference on Software Quality, Reliability and Security (QRS), 2022.
- “TensorFlow: Large-scale machine learning on heterogeneous systems,” 2015, software available from tensorflow.org.
- Y. Zhou, Y. Sun, Y. Tang, Y. Chen, J. Sun, C. M. Poskitt, Y. Liu, and Z. Yang, “Specification-based autonomous driving system testing,” IEEE Transactions on Software Engineering, 2023.
- T. Phan-Minh, F. Howington, T.-S. Chu, M. S. Tomov, R. E. Beaudoin, S. U. Lee, N. Li, C. Dicle, S. Findler, F. Suarez-Ruiz, B. Yang, S. Omari, and E. M. Wolff, “Driveirl: Drive in real life with inverse reinforcement learning,” in 2023 IEEE International Conference on Robotics and Automation (ICRA), 2023.
- Z. Wu, L. Sun, W. Zhan, C. Yang, and M. Tomizuka, “Efficient sampling-based maximum entropy inverse reinforcement learning with application to autonomous driving,” IEEE Robotics and Automation Letters, 2020.
- S. Rosbach, V. James, S. Großjohann, S. Homoceanu, and S. Roth, “Driving with style: Inverse reinforcement learning in general-purpose planning for automated driving,” in 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2019.
- Z. Huang, H. Liu, J. Wu, and C. Lv, “Conditional predictive behavior planning with inverse reinforcement learning for human-like autonomous driving,” IEEE Transactions on Intelligent Transportation Systems, 2023.
- J. Mendez, S. Shivkumar, and E. Eaton, “Lifelong inverse reinforcement learning,” Advances in neural information processing systems, 2018.
- A. K. Tanwani and A. Billard, “Transfer in inverse reinforcement learning for multiple strategies,” in 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, 2013.
- Y. Deng, X. Zheng, T. Zhang, H. Liu, G. Lou, M. Kim, and T. Y. Chen, “A declarative metamorphic testing framework for autonomous driving,” IEEE Transactions on Software Engineering, 2023.
- Y. Tian, K. Pei, S. Jana, and B. Ray, “Deeptest: Automated testing of deep-neural-network-driven autonomous cars,” in Proceedings of the 40th International Conference on Software Engineering, 2018.
- F. Haq, D. Shin, S. Nejati, and L. C. Briand, “Comparing offline and online testing of deep neural networks: An autonomous car case study,” in 2020 IEEE 13th International Conference on Software Testing, Validation and Verification (ICST), 2020.
- B. Gangopadhyay, S. Khastgir, S. Dey, P. Dasgupta, G. Montana, and P. Jennings, “Identification of test cases for automated driving systems using bayesian optimization,” in 2019 IEEE Intelligent Transportation Systems Conference (ITSC), 2019.
- Z. Zhong, G. Kaiser, and B. Ray, “Neural network guided evolutionary fuzzing for finding traffic violations of autonomous vehicles,” IEEE Transactions on Software Engineering, 2022.
- G. Li, Y. Li, S. Jha, T. Tsai, M. Sullivan, S. K. S. Hari, Z. Kalbarczyk, and R. Iyer, “Av-fuzzer: Finding safety violations in autonomous driving systems,” in 2020 IEEE 31st International Symposium on Software Reliability Engineering (ISSRE), 2020.
- C. Lu, Y. Shi, H. Zhang, M. Zhang, T. Wang, T. Yue, and S. Ali, “Learning configurations of operating environment of autonomous vehicles to maximize their collisions,” IEEE Transactions on Software Engineering, vol. 49, no. 1, pp. 384–402, 2022.
- J. Wang, A. Pun, J. Tu, S. Manivasagam, A. Sadat, S. Casas, M. Ren, and R. Urtasun, “Advsim: Generating safety-critical scenarios for self-driving vehicles,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 9909–9918.
- Y. Cao, C. Xiao, D. Yang, J. Fang, R. Yang, M. D. Liu, and B. Li, “Adversarial objects against lidar-based autonomous driving systems,” in arXiv, 2019.
- H. Delecki, M. Itkina, B. Lange, R. Senanayake, and M. J. Kochenderfer, “How do we fail? stress testing perception in autonomous vehicles,” 2022.
- G. Christian, T. Woodlief, and S. Elbaum, “Generating realistic and diverse tests for lidar-based perception systems,” in 2023 IEEE/ACM 45th International Conference on Software Engineering (ICSE).
- A. Guo, Y. Feng, and Z. Chen, “Lirtest: augmenting lidar point clouds for automated testing of autonomous driving systems,” in Proceedings of the 31st ACM SIGSOFT International Symposium on Software Testing and Analysis, 2022.
- W. Ding, B. Chen, M. Xu, and D. Zhao, “Learning to collide: An adaptive safety-critical scenarios generating method,” in 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2020.
- C. Chen, A. Seff, A. Kornhauser, and J. Xiao, “Deepdriving: Learning affordance for direct perception in autonomous driving,” in Proceedings of the IEEE international conference on computer vision, 2015.
- Q. Song, K. Tan, P. Runeson, and S. Persson, “Critical scenario identification for realistic testing of autonomous driving systems,” Software Quality Journal, 2022.
- S. Kim, M. Liu, J. J. Rhee, Y. Jeon, Y. Kwon, and C. H. Kim, “DriveFuzz: Discovering Autonomous Driving Bugs through Driving Quality-Guided Fuzzing,” in Proceedings of the 2022 ACM SIGSAC Conference on Computer and Communications Security, 2022.