Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
125 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Improving Learnt Local MAPF Policies with Heuristic Search (2403.20300v1)

Published 29 Mar 2024 in cs.MA, cs.AI, and cs.RO

Abstract: Multi-agent path finding (MAPF) is the problem of finding collision-free paths for a team of agents to reach their goal locations. State-of-the-art classical MAPF solvers typically employ heuristic search to find solutions for hundreds of agents but are typically centralized and can struggle to scale when run with short timeouts. Machine learning (ML) approaches that learn policies for each agent are appealing as these could enable decentralized systems and scale well while maintaining good solution quality. Current ML approaches to MAPF have proposed methods that have started to scratch the surface of this potential. However, state-of-the-art ML approaches produce "local" policies that only plan for a single timestep and have poor success rates and scalability. Our main idea is that we can improve a ML local policy by using heuristic search methods on the output probability distribution to resolve deadlocks and enable full horizon planning. We show several model-agnostic ways to use heuristic search with learnt policies that significantly improve the policies' success rates and scalability. To our best knowledge, we demonstrate the first time ML-based MAPF approaches have scaled to high congestion scenarios (e.g. 20% agent density).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (17)
  1. Suboptimal variants of the conflict-based search algorithm for the multi-agent pathfinding problem. In Seventh Annual Symposium on Combinatorial Search.
  2. Don’t Split, Try To Work It Out: Bypassing Conflicts in Multi-Agent Pathfinding. Proceedings of the International Conference on Automated Planning and Scheduling, 25(1): 47–51.
  3. PRIMAL _⁢2_2\_2_ 2: Pathfinding via reinforcement and imitation multi-agent learning-lifelong. IEEE Robotics and Automation Letters, 6(2): 2666–2673.
  4. Convolutional Neural Network Architectures for Signals Supported on Graphs. IEEE Transactions on Signal Processing, 67(4): 1034–1049.
  5. MAPF-LNS2: Fast Repairing for Multi-Agent Path Finding via Large Neighborhood Search. Proceedings of the AAAI Conference on Artificial Intelligence, 36(9): 10256–10265.
  6. Improved Heuristics for Multi-Agent Path Finding with Conflict-Based Search. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI-19, 442–449. International Joint Conferences on Artificial Intelligence Organization.
  7. EECBS: A bounded-suboptimal search for multi-agent path finding. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), 12353–12362.
  8. Lifelong Multi-Agent Path Finding in Large-Scale Warehouses. In Proceedings of the 19th International Conference on Autonomous Agents and MultiAgent Systems, AAMAS ’20, 1898–1900. Richland, SC: International Foundation for Autonomous Agents and Multiagent Systems.
  9. Graph Neural Networks for Decentralized Multi-Robot Path Planning. arXiv:1912.06095.
  10. Message-Aware Graph Attention Networks for Large-Scale Multi-Robot Path Planning. arXiv:2011.13219.
  11. Okumura, K. 2022. LaCAM: Search-Based Algorithm for Quick Multi-Agent Pathfinding. arXiv:2211.13432.
  12. Priority inheritance with backtracking for iterative multi-agent path finding. Artificial Intelligence, 310: 103752.
  13. Primal: Pathfinding via reinforcement and imitation multi-agent learning. IEEE Robotics and Automation Letters, 4(3): 2378–2385.
  14. Conflict-based search for optimal multi-agent pathfinding. Artificial Intelligence, 219: 40–66.
  15. The increasing cost tree search for optimal multi-agent pathfinding. Artificial Intelligence, 195: 470–495.
  16. Multi-Agent Pathfinding: Definitions, Variants, and Benchmarks. Symposium on Combinatorial Search (SoCS), 151–158.
  17. SCRIMP: Scalable Communication for Reinforcement- and Imitation-Learning-Based Multi-Agent Pathfinding. arXiv:2303.00605.
Citations (4)

Summary

We haven't generated a summary for this paper yet.