SECRM-2D: RL-Based Efficient and Comfortable Route-Following Autonomous Driving with Analytic Safety Guarantees (2407.16857v1)
Abstract: Over the last decade, there has been increasing interest in autonomous driving systems. Reinforcement Learning (RL) shows great promise for training autonomous driving controllers, being able to directly optimize a combination of criteria such as efficiency comfort, and stability. However, RL- based controllers typically offer no safety guarantees, making their readiness for real deployment questionable. In this paper, we propose SECRM-2D (the Safe, Efficient and Comfortable RL- based driving Model with Lane-Changing), an RL autonomous driving controller (both longitudinal and lateral) that balances optimization of efficiency and comfort and follows a fixed route, while being subject to hard analytic safety constraints. The aforementioned safety constraints are derived from the criterion that the follower vehicle must have sufficient headway to be able to avoid a crash if the leader vehicle brakes suddenly. We evaluate SECRM-2D against several learning and non-learning baselines in simulated test scenarios, including freeway driving, exiting, merging, and emergency braking. Our results confirm that representative previously-published RL AV controllers may crash in both training and testing, even if they are optimizing a safety objective. By contrast, our controller SECRM-2D is successful in avoiding crashes during both training and testing, improves over the baselines in measures of efficiency and comfort, and is more faithful in following the prescribed route. In addition, we achieve a good theoretical understanding of the longitudinal steady-state of a collection of SECRM-2D vehicles.
- V. Mnih, K. Kavukcuoglu, D. Silver, A. A. Rusu, J. Veness, M. G. Bellemare, A. Graves, M. Riedmiller, A. K. Fidjeland, G. Ostrovski, S. Petersen, C. Beattie, A. Sadik, I. Antonoglou, H. King, D. Kumaran, D. Wierstra, S. Legg, and D. Hassabis, “Human-level control through deep reinforcement learning,” Nature, vol. 518, no. 7540, pp. 529–533, 2015. [Online]. Available: https://doi.org/10.1038/nature14236
- N. Lazic, C. Boutilier, T. Lu, E. Wong, B. Roy, M. Ryu, and G. Imwalle, “Data center cooling using model-predictive control,” Advances in Neural Information Processing Systems, vol. 31, 2018.
- G. Li, Y. Yang, S. Li, X. Qu, N. Lyu, and S. E. Li, “Decision making of autonomous vehicles in lane change scenarios: Deep reinforcement learning approaches with risk awareness,” Transportation Research Part C: Emerging Technologies, vol. 134, p. 103452, 2022.
- O. Elsamadisy, T. Shi, I. Smirnov, and B. Abdulhai, “Safe, efficient and comfortable reinforcement-learning-based car-following for AVs with analytic safety guarantee and dynamic target speed,” Transportation Research Record, vol. 2678(1), pp. 643–661, 2024.
- S. Zhang, H. Peng, S. Nageshrao, and E. Tseng, “Discretionary lane change decision making using reinforcement learning with model-based exploration,” in 2019 18th IEEE International Conference On Machine Learning And Applications (ICMLA). IEEE, 2019, pp. 844–850.
- C. Xi, T. Shi, Y. Wu, and L. Sun, “Efficient motion planning for automated lane change based on imitation learning and mixed-integer optimization,” in 2020 IEEE 23rd International Conference on Intelligent Transportation Systems (ITSC). IEEE, 2020, pp. 1–6.
- T. Pan, W. H. Lam, A. Sumalee, and R. Zhong, “Modeling the impacts of mandatory and discretionary lane-changing maneuvers,” Transportation research part C: emerging technologies, vol. 68, pp. 403–424, 2016.
- T. Shi, P. Wang, X. Cheng, C.-Y. Chan, and D. Huang, “Driving decision and control for automated lane change behavior based on deep reinforcement learning,” in 2019 IEEE intelligent transportation systems conference (ITSC). IEEE, 2019, pp. 2895–2900.
- C.-J. Hoel, K. Driggs-Campbell, K. Wolff, L. Laine, and M. J. Kochenderfer, “Combining planning and deep reinforcement learning in tactical decision making for autonomous driving,” IEEE transactions on intelligent vehicles, vol. 5, no. 2, pp. 294–305, 2019.
- Z. Cao, D. Yang, S. Xu, H. Peng, B. Li, S. Feng, and D. Zhao, “Highway exiting planner for automated vehicles using reinforcement learning,” IEEE Transactions on Intelligent Transportation Systems, vol. 22, no. 2, pp. 990–1000, 2020.
- F. Ye, X. Cheng, P. Wang, C.-Y. Chan, and J. Zhang, “Automated lane change strategy using proximal policy optimization-based deep reinforcement learning,” in 2020 IEEE Intelligent Vehicles Symposium (IV). IEEE, 2020, pp. 1746–1752.
- S. Udatha, Y. Lyu, and J. Dolan, “Reinforcement learning with probabilistically safe control barrier functions for ramp merging,” in 2023 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 2023, pp. 5625–5630.
- B. Vanholme, D. Gruyer, B. Lusetti, S. Glaser, and S. Mammar, “Highly automated driving on highways based on legal safety,” IEEE Transactions on Intelligent Transportation Systems, vol. 14, no. 1, pp. 333–347, 2013.
- M. Zhu, Y. Wang, Z. Pu, J. Hu, X. Wang, and R. Ke, “Safe, efficient, and comfortable velocity control based on reinforcement learning for autonomous driving,” Transportation Research Part C: Emerging Technologies, vol. 117, p. 102662, 2020.
- Y.-T. Yen, J.-J. Chou, C.-S. Shih, C.-W. Chen, and P.-K. Tsung, “Proactive car-following using deep-reinforcement learning,” in 2020 IEEE 23rd International Conference on Intelligent Transportation Systems (ITSC). IEEE, 2020, pp. 1–6.
- T. Shi, Y. Ai, O. ElSamadisy, and B. Abdulhai, “Bilateral deep reinforcement learning approach for better-than-human car-following,” in 2022 IEEE 25th International Conference on Intelligent Transportation Systems (ITSC). IEEE, 2022, pp. 3986–3992.
- P. Gipps, “A behavioural car-following model for computer simulation,” Transportation Research Part B: Methodological, vol. 15, no. 2, pp. 105–111, 1981.
- W. Zhao, T. He, R. Chen, T. Wei, and C. Liu, “State-wise safe reinforcement learning: A survey,” arXiv preprint arXiv:2302.03122, 2023.
- T. P. Lillicrap, J. J. Hunt, A. Pritzel, N. Heess, T. Erez, Y. Tassa, D. Silver, and D. Wierstra, “Continuous control with deep reinforcement learning,” in 4th International Conference on Learning Representations, ICLR 2016, San Juan, Puerto Rico, May 2-4, 2016, Conference Track Proceedings, Y. Bengio and Y. LeCun, Eds., 2016. [Online]. Available: http://arxiv.org/abs/1509.02971
- P. Á. López, M. Behrisch, L. Bieker-Walz, J. Erdmann, Y.-P. Flötteröd, R. Hilbrich, L. Lücken, J. Rummel, P. Wagner, and E. WieBner, “Microscopic traffic simulation using SUMO,” 2018 21st International Conference on Intelligent Transportation Systems (ITSC), pp. 2575–2582, 2018.
- Treiber, Hennecke, and Helbing, “Congested traffic states in empirical observations and microscopic simulations,” Physical Review E: Statistical Physics, Plasmas, Fluids, and Related Interdisciplinary Topics, vol. 62 2 Pt A, pp. 1805–24, 2000.
- A. Kesting, M. Treiber, and D. Helbing, “General lane-changing model MOBIL for car-following models,” Transportation Research Record, vol. 1999(1), pp. 86–94, 2007.
- J. Erdmann, “SUMO’s lane-changing model,” in Modeling Mobility with Open Data: 2nd SUMO Conference 2014 Berlin, Germany, May 15-16, 2014. Springer, 2015, pp. 105–123.
- T. Shi, “SECRM-2D demo,” 2023, https://youtu.be/J5UyJXGAGX0, Last accessed May 20, 2023.