Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
166 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Bigraph Matching Weighted with Learnt Incentive Function for Multi-Robot Task Allocation (2403.07131v1)

Published 11 Mar 2024 in cs.AI and cs.MA

Abstract: Most real-world Multi-Robot Task Allocation (MRTA) problems require fast and efficient decision-making, which is often achieved using heuristics-aided methods such as genetic algorithms, auction-based methods, and bipartite graph matching methods. These methods often assume a form that lends better explainability compared to an end-to-end (learnt) neural network based policy for MRTA. However, deriving suitable heuristics can be tedious, risky and in some cases impractical if problems are too complex. This raises the question: can these heuristics be learned? To this end, this paper particularly develops a Graph Reinforcement Learning (GRL) framework to learn the heuristics or incentives for a bipartite graph matching approach to MRTA. Specifically a Capsule Attention policy model is used to learn how to weight task/robot pairings (edges) in the bipartite graph that connects the set of tasks to the set of robots. The original capsule attention network architecture is fundamentally modified by adding encoding of robots' state graph, and two Multihead Attention based decoders whose output are used to construct a LogNormal distribution matrix from which positive bigraph weights can be drawn. The performance of this new bigraph matching approach augmented with a GRL-derived incentive is found to be at par with the original bigraph matching approach that used expert-specified heuristics, with the former offering notable robustness benefits. During training, the learned incentive policy is found to get initially closer to the expert-specified incentive and then slightly deviate from its trend.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (32)
  1. Y. Meng and J. Gan, “A distributed swarm intelligence based algorithm for a cooperative multi-robot construction task,” in 2008 IEEE Swarm Intelligence Symposium.   IEEE, 2008, pp. 1–6.
  2. P. Ghassemi and S. Chowdhury, “Multi-robot task allocation in disaster response: Addressing dynamic tasks with deadlines and robots with range and payload constraints,” Robotics and Autonomous Systems, p. 103905, 2021.
  3. Y. Huang, Y. Zhang, and H. Xiao, “Multi-robot system task allocation mechanism for smart factory,” in 2019 IEEE 8th Joint International Information Technology and Artificial Intelligence Conference (ITAIC), 2019, pp. 587–591.
  4. F. Xue, H. Tang, Q. Su, and T. Li, “Task allocation of intelligent warehouse picking system based on multi-robot coalition,” KSII Transactions on Internet and Information Systems (TIIS), vol. 13, no. 7, pp. 3566–3582, 2019.
  5. Y. Cao, S. Wang, and J. Li, “The optimization model of ride-sharing route for ride hailing considering both system optimization and user fairness,” Sustainability, vol. 13, no. 2, 2021. [Online]. Available: https://www.mdpi.com/2071-1050/13/2/902
  6. A. Farinelli, A. Rogers, A. Petcu, and N. R. Jennings, “Decentralised coordination of low-power embedded devices using the max-sum algorithm,” 2008.
  7. P. Ghassemi, D. DePauw, and S. Chowdhury, “Decentralized dynamic task allocation in swarm robotic systems for disaster response: Extended abstract,” in 2019 International Symposium on Multi-Robot and Multi-Agent Systems (MRS), 2019, pp. 83–85.
  8. M. B. Dias, R. Zlot, N. Kalra, and A. Stentz, “Market-based multirobot coordination: A survey and analysis,” Proceedings of the IEEE, vol. 94, no. 7, pp. 1257–1270, 2006.
  9. N. Mazyavkina, S. Sviridov, S. Ivanov, and E. Burnaev, “Reinforcement learning for combinatorial optimization: A survey,” Computers & Operations Research, vol. 134, p. 105400, 2021.
  10. S. Paul, W. Li, B. Smyth, Y. Chen, Y. Gel, and S. Chowdhury, “Efficient planning of multi-robot collective transport using graph reinforcement learning with higher order topological abstraction,” arXiv preprint arXiv:2303.08933, 2023.
  11. K. Jose and D. K. Pratihar, “Task allocation and collision-free path planning of centralized multi-robots system for industrial plant inspection using heuristic methods,” Robotics and Autonomous Systems, vol. 80, pp. 34–42, 2016.
  12. D. Cattaruzza, N. Absi, and D. Feillet, “Vehicle routing problems with multiple trips,” 4OR, vol. 14, no. 3, pp. 223–259, 2016.
  13. E. Schneider, E. I. Sklar, S. Parsons, and A. T. Özgelen, “Auction-based task allocation for multi-robot teams in dynamic environments,” in Conference Towards Autonomous Robotic Systems.   Springer, 2015, pp. 246–257.
  14. M. Otte, M. J. Kuhlman, and D. Sofge, “Auctions for multi-robot task allocation in communication limited environments,” Autonomous Robots, vol. 44, no. 3, pp. 547–584, 2020.
  15. H. Mitiche, D. Boughaci, and M. Gini, “Iterated local search for time-extended multi-robot task allocation with spatio-temporal and capacity constraints,” Journal of Intelligent Systems, 2019.
  16. P. Vansteenwegen, W. Souffriau, G. Vanden Berghe, and D. Van Oudheusden, “Iterated local search for the team orienteering problem with time windows,” Computers & Operations Research, vol. 36, no. 12, pp. 3281–3290, 2009, new developments on hub location. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S030505480900080X
  17. S. Ismail and L. Sun, “Decentralized hungarian-based approach for fast and scalable task allocation,” in 2017 International Conference on Unmanned Aircraft Systems (ICUAS).   IEEE, 2017, pp. 23–28.
  18. W. Kool, H. Van Hoof, and M. Welling, “Attention, learn to solve routing problems!” in 7th International Conference on Learning Representations, ICLR 2019, 2019.
  19. Y. Kaempfer and L. Wolf, “Learning the multiple traveling salesmen problem with permutation invariant pooling networks,” ArXiv, vol. abs/1803.09621, 2018.
  20. E. Khalil, H. Dai, Y. Zhang, B. Dilkina, and L. Song, “Learning combinatorial optimization algorithms over graphs,” in Advances in Neural Information Processing Systems, 2017, pp. 6348–6358.
  21. S. Paul and S. Chowdhury, “A scalable graph learning approach to capacitated vehicle routing problem using capsule networks and attention mechanism,” in International Design Engineering Technical Conferences and Computers and Information in Engineering Conference, vol. 86236.   American Society of Mechanical Engineers, 2022, p. V03BT03A045.
  22. E. V. Tolstaya, J. Paulos, V. R. Kumar, and A. Ribeiro, “Multi-robot coverage and exploration using spatial graph neural networks,” ArXiv, vol. abs/2011.01119, 2020.
  23. S. Paul, P. Ghassemi, and S. Chowdhury, “Learning scalable policies over graphs for multi-robot task allocation using capsule attention networks,” in 2022 International Conference on Robotics and Automation (ICRA), 2022, pp. 8815–8822.
  24. R. A. Jacob, S. Paul, W. Li, S. Chowdhury, Y. R. Gel, and J. Zhang, “Reconfiguring unbalanced distribution networks using reinforcement learning over graphs,” in 2022 IEEE Texas Power and Energy Conference (TPEC), 2022, pp. 1–6.
  25. S. Paul and S. Chowdhury, “A graph-based reinforcement learning framework for urban air mobility fleet scheduling,” in AIAA AVIATION 2022 Forum, 2022, p. 3911.
  26. P. KrisshnaKumar, J. Witter, S. Paul, H. Cho, K. Dantu, and S. Chowdhury, “Fast decision support for air traffic management at urban air mobility vertiports using graph learning,” in 2023 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).   IEEE, 2023, pp. 1580–1585.
  27. P. K. Kumar, J. Witter, S. Paul, K. Dantu, and S. Chowdhury, “Graph learning based decision support for multi-aircraft take-off and landing at urban air mobility vertiports,” in AIAA SCITECH 2023 Forum, 2023, p. 1848.
  28. H. W. Kuhn, “The hungarian method for the assignment problem,” Naval Research Logistics Quarterly, vol. 2, no. 1-2, pp. 83–97, 1955.
  29. J. E. Hopcroft and R. M. Karp, “An n5̂/2 algorithm for maximum matchings in bipartite graphs,” SIAM Journal on Computing, vol. 2, no. 4, pp. 225–231, 1973.
  30. S. Verma and Z. L. Zhang, “Graph capsule convolutional neural networks,” 2018.
  31. J. Schulman, F. Wolski, P. Dhariwal, A. Radford, and O. Klimov, “Proximal policy optimization algorithms,” 2017.
  32. A. Raffin, A. Hill, A. Gleave, A. Kanervisto, M. Ernestus, and N. Dormann, “Stable-baselines3: Reliable reinforcement learning implementations,” Journal of Machine Learning Research, vol. 22, no. 268, pp. 1–8, 2021. [Online]. Available: http://jmlr.org/papers/v22/20-1364.html

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com