Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Optimizing UAV-UGV Coalition Operations: A Hybrid Clustering and Multi-Agent Reinforcement Learning Approach for Path Planning in Obstructed Environment (2401.01481v1)

Published 3 Jan 2024 in cs.RO and cs.MA

Abstract: One of the most critical applications undertaken by coalitions of Unmanned Aerial Vehicles (UAVs) and Unmanned Ground Vehicles (UGVs) is reaching predefined targets by following the most time-efficient routes while avoiding collisions. Unfortunately, UAVs are hampered by limited battery life, and UGVs face challenges in reachability due to obstacles and elevation variations. Existing literature primarily focuses on one-to-one coalitions, which constrains the efficiency of reaching targets. In this work, we introduce a novel approach for a UAV-UGV coalition with a variable number of vehicles, employing a modified mean-shift clustering algorithm to segment targets into multiple zones. Each vehicle utilizes Multi-agent Deep Deterministic Policy Gradient (MADDPG) and Multi-agent Proximal Policy Optimization (MAPPO), two advanced reinforcement learning algorithms, to form an effective coalition for navigating obstructed environments without collisions. This approach of assigning targets to various circular zones, based on density and range, significantly reduces the time required to reach these targets. Moreover, introducing variability in the number of UAVs and UGVs in a coalition enhances task efficiency by enabling simultaneous multi-target engagement. The results of our experimental evaluation demonstrate that our proposed method substantially surpasses current state-of-the-art techniques, nearly doubling efficiency in terms of target navigation time and task completion rate.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (27)
  1. Unmanned aerial vehicles (uavs): Practical aspects, applications, open challenges, security issues, and future trends, Intelligent Service Robotics (2023). doi:10.1007/s11370-022-00452-4.
  2. S. Caska, A. Gayretli, A survey of uav/ugv collaborative systems (2014) 453–463.
  3. Air-ground spatial crowdsourcing with uav carriers by geometric graph convolutional multi-agent deep reinforcement learning, in: 2023 IEEE 39th International Conference on Data Engineering (ICDE), IEEE, 2023, pp. 1790–1802.
  4. Coordination between unmanned aerial and ground vehicles: A taxonomy and optimization perspective, IEEE transactions on cybernetics 46 (2015). doi:10.1109/TCYB.2015.2418337.
  5. Joint optimization of multi-uav target assignment and path planning based on multi-agent reinforcement learning, IEEE access 7 (2019) 146264–146272.
  6. Iadrl: Imitation augmented deep reinforcement learning enabled ugv-uav coalition for tasking in complex environments, IEEE Access 8 (2020) 102335–102347.
  7. Receding horizon control of autonomous aerial vehicles, in: Proceedings of the 2002 American Control Conference (IEEE Cat. No.CH37301), volume 5, 2002, pp. 3741–3746 vol.5. doi:10.1109/ACC.2002.1024509.
  8. Uav cooperative multiple task assignments using genetic algorithms, in: Proceedings of the 2005, American Control Conference, 2005., IEEE, 2005, pp. 2989–2994.
  9. Particle swarm optimization for resource allocation in uav cooperative control, in: AIAA Guidance, Navigation, and Control Conference and Exhibit, 2004, p. 5250.
  10. L. Babel, Coordinated target assignment and uav path planning with timing constraints, Journal of Intelligent & Robotic Systems 94 (2019) 857–869.
  11. A new method of uavs multi-target task assignment, DEStech Trans. Eng. Technol. Res (2018) 388–394.
  12. C. Yan, X. Xiang, A path planning algorithm for uav based on improved q-learning, in: 2018 2nd international conference on robotics and automation sciences (ICRAS), IEEE, 2018, pp. 1–5.
  13. B. Li, Y. Wu, Path planning for uav ground target tracking via deep reinforcement learning, IEEE Access 8 (2020) 29064–29074.
  14. S. Josef, A. Degani, Deep reinforcement learning for safe local planning of a ground vehicle in unknown rough terrain, IEEE Robotics and Automation Letters 5 (2020) 6748–6755.
  15. Human-level control through deep reinforcement learning, nature 518 (2015) 529–533.
  16. Continuous control with deep reinforcement learning, arXiv preprint arXiv:1509.02971 (2015).
  17. Multi-agent actor-critic for mixed cooperative-competitive environments, Advances in neural information processing systems 30 (2017).
  18. Proximal policy optimization algorithms, arXiv preprint arXiv:1707.06347 (2017).
  19. Meta proximal policy optimization for cooperative multi-agent continuous control, in: 2022 International Joint Conference on Neural Networks (IJCNN), IEEE, 2022, pp. 1–8.
  20. The surprising effectiveness of ppo in cooperative multi-agent games, Advances in Neural Information Processing Systems 35 (2022) 24611–24624.
  21. Coordinated proximal policy optimization, Advances in Neural Information Processing Systems 34 (2021) 26437–26448.
  22. Cooperative uav trajectory design for disaster area emergency communications: A multi-agent ppo method, IEEE Internet of Things Journal (2023) 1–1. doi:10.1109/JIOT.2023.3320796.
  23. Trans-utpa: Pso and maddpg based multi-uavs trajectory planning algorithm for emergency communication, Frontiers in Neurorobotics 16 (2023). URL: https://www.frontiersin.org/articles/10.3389/fnbot.2022.1076338. doi:10.3389/fnbot.2022.1076338.
  24. A clustering-based coverage path planning method for autonomous heterogeneous uavs, IEEE Transactions on Intelligent Transportation Systems 23 (2022a) 25546–25556. doi:10.1109/TITS.2021.3066240.
  25. An adaptive clustering-based algorithm for automatic path planning of heterogeneous uavs, IEEE Transactions on Intelligent Transportation Systems 23 (2022b) 16842–16853. doi:10.1109/TITS.2021.3131473.
  26. Terra: A path planning algorithm for cooperative ugv–uav exploration, Engineering Applications of Artificial Intelligence 78 (2019) 260–272.
  27. Continuous deep hierarchical reinforcement learning for ground-air swarm shepherding, arXiv preprint arXiv:2004.11543 (2020).
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Shamyo Brotee (1 paper)
  2. Farhan Kabir (1 paper)
  3. Md. Abdur Razzaque (2 papers)
  4. Palash Roy (1 paper)
  5. Md. Mamun-Or-Rashid (4 papers)
  6. Md. Rafiul Hassan (2 papers)
  7. Mohammad Mehedi Hassan (6 papers)
Citations (5)

Summary

We haven't generated a summary for this paper yet.