Delay-Aware Multi-Agent Reinforcement Learning for Cooperative Adaptive Cruise Control with Model-based Stability Enhancement (2404.15696v2)
Abstract: Cooperative Adaptive Cruise Control (CACC) represents a quintessential control strategy for orchestrating vehicular platoon movement within Connected and Automated Vehicle (CAV) systems, significantly enhancing traffic efficiency and reducing energy consumption. In recent years, the data-driven methods, such as reinforcement learning (RL), have been employed to address this task due to their significant advantages in terms of efficiency and flexibility. However, the delay issue, which often arises in real-world CACC systems, is rarely taken into account by current RL-based approaches. To tackle this problem, we propose a Delay-Aware Multi-Agent Reinforcement Learning (DAMARL) framework aimed at achieving safe and stable control for CACC. We model the entire decision-making process using a Multi-Agent Delay-Aware Markov Decision Process (MADA-MDP) and develop a centralized training with decentralized execution (CTDE) MARL framework for distributed control of CACC platoons. An attention mechanism-integrated policy network is introduced to enhance the performance of CAV communication and decision-making. Additionally, a velocity optimization model-based action filter is incorporated to further ensure the stability of the platoon. Experimental results across various delay conditions and platoon sizes demonstrate that our approach consistently outperforms baseline methods in terms of platoon safety, stability and overall performance.
- V. Lesch, M. Breitbach, M. Segata, C. Becker, S. Kounev, and C. Krupitzer, “An overview on approaches for coordination of platoons,” IEEE Transactions on Intelligent Transportation Systems, vol. 23, no. 8, pp. 10 049–10 065, 2021.
- Z. Wang, G. Wu, P. Hao, K. Boriboonsomsin, and M. Barth, “Developing a platoon-wide eco-cooperative adaptive cruise control (cacc) system,” in 2017 ieee intelligent vehicles symposium (iv). IEEE, 2017, pp. 1256–1261.
- Z. Wang, G. Wu, and M. J. Barth, “A review on cooperative adaptive cruise control (cacc) systems: Architectures, controls, and applications,” in 2018 21st International Conference on Intelligent Transportation Systems (ITSC). IEEE, 2018, pp. 2884–2891.
- C. Massera Filho, M. H. Terra, and D. F. Wolf, “Safe optimization of highway traffic with robust model predictive control-based cooperative adaptive cruise control,” IEEE Transactions on Intelligent Transportation Systems, vol. 18, no. 11, pp. 3193–3203, 2017.
- C. Wu, A. M. Bayen, and A. Mehta, “Stabilizing traffic with autonomous vehicles,” in Proceedings of 2018 IEEE International Conference on Robotics and Automation, 2018, pp. 6012–6018.
- M. Wang, “Infrastructure assisted adaptive driving to stabilise heterogeneous vehicle strings,” Transportation Research Part C: Emerging Technologies, vol. 91, pp. 276–295, 2018.
- S. Feng, Y. Zhang, S. E. Li, Z. Cao, H. X. Liu, and L. Li, “String stability for vehicular platoon control: Definitions and analysis methods,” Annual Reviews in Control, vol. 47, pp. 81–97, 2019.
- W. Gao, Z.-P. Jiang, and K. Ozbay, “Data-driven adaptive optimal control of connected vehicles,” IEEE Transactions on Intelligent Transportation Systems, vol. 18, no. 5, pp. 1122–1133, 2016.
- A. Peake, J. McCalmon, B. Raiford, T. Liu, and S. Alqahtani, “Multi-agent reinforcement learning for cooperative adaptive cruise control,” in Proceedings of 2020 IEEE 32nd International Conference on Tools with Artificial Intelligence, 2020, pp. 15–22.
- T. Chu and U. Kalabić, “Model-based deep reinforcement learning for cacc in mixed-autonomy vehicle platoon,” in Proceedings of 2019 IEEE 58th Conference on Decision and Control, 2019, pp. 4079–4084.
- L. Lei, T. Liu, K. Zheng, and L. Hanzo, “Deep reinforcement learning aided platoon control relying on v2x information,” IEEE Transactions on Vehicular Technology, vol. 71, no. 6, pp. 5811–5826, 2022.
- L. Jiang, Y. Xie, N. G. Evans, X. Wen, T. Li, and D. Chen, “Reinforcement learning based cooperative longitudinal control for reducing traffic oscillations and improving platoon stability,” Transportation Research Part C: Emerging Technologies, vol. 141, p. 103744, 2022.
- T. Liu, L. Lei, K. Zheng, and K. Zhang, “Autonomous platoon control with integrated deep reinforcement learning and dynamic programming,” IEEE Internet of Things Journal, vol. 10, no. 6, pp. 5476–5489, 2022.
- J. Liu, P. Hang, X. Qi, J. Wang, and J. Sun, “Mtd-gpt: A multi-task decision-making gpt model for autonomous driving at unsignalized intersections,” in 2023 IEEE 26th International Conference on Intelligent Transportation Systems (ITSC). IEEE, 2023, pp. 5154–5161.
- J. Liu, D. Zhou, P. Hang, Y. Ni, and J. Sun, “Towards socially responsive autonomous vehicles: A reinforcement learning framework with driving priors and coordination awareness,” IEEE Transactions on Intelligent Vehicles, vol. 9, no. 1, pp. 827–838, 2024.
- J. Wang, X. Xu, D. Liu, Z. Sun, and Q. Chen, “Self-learning cruise control using kernel-based least squares policy iteration,” IEEE Transactions on Control Systems Technology, vol. 22, no. 3, pp. 1078–1087, 2013.
- Z. Li, T. Chu, I. V. Kolmanovsky, and X. Yin, “Training drift counteraction optimal control policies using reinforcement learning: An adaptive cruise control example,” IEEE transactions on intelligent transportation systems, vol. 19, no. 9, pp. 2903–2912, 2017.
- C. Desjardins and B. Chaib-Draa, “Cooperative adaptive cruise control: A reinforcement learning approach,” IEEE Transactions on intelligent transportation systems, vol. 12, no. 4, pp. 1248–1260, 2011.
- Z. Wang, S. Jin, L. Liu, C. Fang, M. Li, and S. Guo, “Design of intelligent connected cruise control with vehicle-to-vehicle communication delays,” IEEE Transactions on Vehicular Technology, vol. 71, no. 8, pp. 9011–9025, 2022.
- B. Chen, M. Xu, Z. Liu, L. Li, and D. Zhao, “Delay-aware multi-agent reinforcement learning for cooperative and competitive environments,” arXiv preprint arXiv:2005.05441, 2020.
- J. Liu, J. Sun, and X. Qi, “Optimal placement of charging stations in road networks: A reinforcement learning approach with attention mechanism,” Applied Sciences, vol. 13, no. 14, p. 8473, 2023.
- E. Schuitema, L. Buşoniu, R. Babuška, and P. Jonker, “Control delay in reinforcement learning for real-time dynamic systems: A memoryless approach,” in 2010 IEEE/RSJ international conference on intelligent robots and systems. IEEE, 2010, pp. 3226–3231.
- K. V. Katsikopoulos and S. E. Engelbrecht, “Markov decision processes with delays and asynchronous cost collection,” IEEE transactions on automatic control, vol. 48, no. 4, pp. 568–574, 2003.
- H. Zhu, Y. Zhou, X. Luo, and H. Zhou, “Joint control of power, beamwidth, and spacing for platoon-based vehicular cyber-physical systems,” IEEE Transactions on Vehicular Technology, vol. 71, no. 8, pp. 8615–8629, 2022.
- J. Liu, P. Hang, X. Na, C. Huang, and J. Sun, “Cooperative decision-making for cavs at unsignalized intersections: A marl approach with attention and hierarchical game priors,” Authorea Preprints, 2023.
- H. Xing, J. Ploeg, and H. Nijmeijer, “Compensation of communication delays in a cooperative acc system,” IEEE Transactions on Vehicular Technology, vol. 69, no. 2, pp. 1177–1189, 2019.
- J. Liu, Y. Zhou, and L. Liu, “Communication delay-aware cooperative adaptive cruise control with dynamic network topologies-a convergence of communication and control,” Digital Communications and Networks, 2023.
- R. Lowe, Y. I. Wu, A. Tamar, J. Harb, O. Pieter Abbeel, and I. Mordatch, “Multi-agent actor-critic for mixed cooperative-competitive environments,” Advances in neural information processing systems, vol. 30, 2017.
- K. Zhang, Z. Yang, H. Liu, T. Zhang, and T. Basar, “Fully decentralized multi-agent reinforcement learning with networked agents,” in Proceedings of International Conference on Machine Learning. PMLR, 2018, pp. 5872–5881.
- T. Chu, S. Chinchali, and S. Katti, “Multi-agent reinforcement learning for networked system control,” in Proceedings of International Conference on Learning Representations, 2020. [Online]. Available: https://openreview.net/forum?id=Syx7A3NFvH
- Jiaqi Liu (102 papers)
- Ziran Wang (49 papers)
- Peng Hang (34 papers)
- Jian Sun (415 papers)