Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
166 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning Neural Traffic Rules (2312.01498v1)

Published 3 Dec 2023 in cs.RO

Abstract: Extensive research has been devoted to the field of multi-agent navigation. Recently, there has been remarkable progress attributed to the emergence of learning-based techniques with substantially elevated intelligence and realism. Nonetheless, prevailing learned models face limitations in terms of scalability and effectiveness, primarily due to their agent-centric nature, i.e., the learned neural policy is individually deployed on each agent. Inspired by the efficiency observed in real-world traffic networks, we present an environment-centric navigation policy. Our method learns a set of traffic rules to coordinate a vast group of unintelligent agents that possess only basic collision-avoidance capabilities. Our method segments the environment into distinct blocks and parameterizes the traffic rule using a Graph Recurrent Neural Network (GRNN) over the block network. Each GRNN node is trained to modulate the velocities of agents as they traverse through. Using either Imitation Learning (IL) or Reinforcement Learning (RL) schemes, we demonstrate the efficacy of our neural traffic rules in resolving agent congestion, closely resembling real-world traffic regulations. Our method handles up to $240$ agents at real-time and generalizes across diverse agent and environment configurations.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (40)
  1. Kaveh Azadeh, René De Koster and Debjit Roy “Robotized and automated warehouse systems: Review and recent developments” In Transportation Science 53.4 INFORMS, 2019, pp. 917–945
  2. “A survey of autonomous driving: Common practices and emerging technologies” In IEEE access 8 IEEE, 2020, pp. 58443–58469
  3. Bhagya Nathali Silva, Murad Khan and Kijun Han “Towards sustainable smart cities: A review of trends, architectures, components, and open challenges in smart cities” In Sustainable cities and society 38 Elsevier, 2018, pp. 697–713
  4. Boris De Wilde, Adriaan W Ter Mors and Cees Witteveen “Push and rotate: cooperative multi-agent path planning” In Proceedings of the 2013 international conference on Autonomous agents and multi-agent systems, 2013, pp. 87–94
  5. “Structure and intractability of optimal multi-robot path planning on graphs” In Proceedings of the AAAI Conference on Artificial Intelligence 27.1, 2013, pp. 1443–1449
  6. “Conflict-based search for optimal multi-agent pathfinding” In Artificial Intelligence 219 Elsevier, 2015, pp. 40–66
  7. Jur Berg, Ming Lin and Dinesh Manocha “Reciprocal velocity obstacles for real-time multi-agent navigation” In 2008 IEEE international conference on robotics and automation, 2008, pp. 1928–1935 Ieee
  8. “Implicit crowds: Optimization integrator for robust crowd simulation” In ACM Transactions on Graphics (TOG) 36.4 ACM New York, NY, USA, 2017, pp. 1–13
  9. “Distributed multi-robot collision avoidance via deep reinforcement learning for navigation in complex scenarios” In The International Journal of Robotics Research 39.7 SAGE Publications Sage UK: London, England, 2020, pp. 856–892
  10. “Decentralized, unlabeled multi-agent navigation in obstacle-rich environments using graph neural networks” In 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2021, pp. 8936–8943 IEEE
  11. “A review of artificial intelligence applied to path planning in UAV swarms” In Neural Computing and Applications Springer, 2022, pp. 1–18
  12. “Sambot: A self-assembly modular robot for swarm robot” In 2010 IEEE International Conference on Robotics and Automation, 2010, pp. 66–71 IEEE
  13. Michael Rubenstein, Christian Ahler and Radhika Nagpal “Kilobot: A low cost scalable robot system for collective behaviors” In 2012 IEEE international conference on robotics and automation, 2012, pp. 3293–3298 IEEE
  14. “Graph neural networks for decentralized multi-robot path planning” In 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2020, pp. 11785–11792 IEEE
  15. Jingjin Yu “Average case constant factor time and distance optimal multi-robot path planning in well-connected environments” In Autonomous Robots 44.3-4 Springer, 2020, pp. 469–483
  16. “Motion planning for unlabeled discs with optimality guarantees” In arXiv preprint arXiv:1504.05218, 2015
  17. “A review on crowd simulation and modeling” In Graphical Models 111 Elsevier, 2020, pp. 101081
  18. “Dynamic group behaviors for interactive crowd simulation” In Proceedings of the ACM SIGGRAPH/Eurographics Symposium on Computer Animation, 2016, pp. 139–147
  19. Liang He, Zherong Pan and Dinesh Manocha “Real-Time Decentralized Navigation of Nonholonomic Agents Using Shifted Yielding Areas” In 2023 IEEE International Conference on Robotics and Automation (ICRA), 2023, pp. 3656–3662 DOI: 10.1109/ICRA48891.2023.10160902
  20. “Situation agents: agent-based externalized steering logic” In Computer Animation and Virtual Worlds 21.3-4 Wiley Online Library, 2010, pp. 267–276
  21. Taoan Huang, Sven Koenig and Bistra Dilkina “Learning to resolve conflicts for multi-agent path finding with conflict-based search” In Proceedings of the AAAI Conference on Artificial Intelligence 35.13, 2021, pp. 11246–11253
  22. Shuai D. Han and Jingjin Yu “DDM: Fast Near-Optimal Multi-Robot Path Planning Using Diversified-Path and Optimal Sub-Problem Solution Database Heuristics” Note: presented at ICRA 2020 In IEEE Robotics and Automation Letters 5.2, 2020, pp. 1350–1357 DOI: 10.1109/LRA.2020.2967326
  23. “Deepmnavigate: Deep reinforced multi-robot navigation unifying local & global collision avoidance” In 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2020, pp. 6952–6959 IEEE
  24. “The emergence of adversarial communication in multi-agent reinforcement learning” In Conference on Robot Learning, 2021, pp. 1394–1414 PMLR
  25. “Differentiable Learning of Scalable Multi-Agent Navigation Policies” In IEEE Robotics and Automation Letters 8.4 IEEE, 2023, pp. 2229–2236
  26. “Realistic data-driven traffic flow animation using texture synthesis” In IEEE transactions on visualization and computer graphics 24.2 IEEE, 2017, pp. 1167–1178
  27. “Trafficpredict: Trajectory prediction for heterogeneous traffic-agents” In Proceedings of the AAAI conference on artificial intelligence 33.01, 2019, pp. 6120–6127
  28. “Trafficsim: Learning to simulate realistic multi-agent behaviors” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 10400–10409
  29. “A survey on reinforcement learning models and algorithms for traffic signal control” In ACM Computing Surveys (CSUR) 50.3 ACM New York, NY, USA, 2017, pp. 1–38
  30. “Recognizing text-based traffic signs” In IEEE Transactions on Intelligent Transportation Systems 16.3 IEEE, 2014, pp. 1360–1369
  31. Ruben J Franklin “Traffic signal violation detection using artificial intelligence and deep learning” In 2020 5th international conference on communication and electronics systems (ICCES), 2020, pp. 839–844 IEEE
  32. Chris Lee, Bruce Hellinga and Frank Saccomanno “Real-time crash prediction model for application to crash prevention in freeway traffic” In Transportation Research Record 1840.1 SAGE Publications Sage CA: Los Angeles, CA, 2003, pp. 67–77
  33. Luana Ruiz, Fernando Gama and Alejandro Ribeiro “Gated graph recurrent neural networks” In IEEE Transactions on Signal Processing 68 IEEE, 2020, pp. 6303–6318
  34. Will Hamilton, Zhitao Ying and Jure Leskovec “Inductive representation learning on large graphs” In Advances in neural information processing systems 30, 2017
  35. Tomás Lozano-Pérez and Michael A Wesley “An algorithm for planning collision-free paths among polyhedral obstacles” In Communications of the ACM 22.10 ACM New York, NY, USA, 1979, pp. 560–570
  36. “Benchmarks for reinforcement learning in mixed-autonomy traffic” In Conference on robot learning, 2018, pp. 399–409 PMLR
  37. Joshua Hare “Dealing with sparse rewards in reinforcement learning” In arXiv preprint arXiv:1910.09281, 2019
  38. Stéphane Ross, Geoffrey Gordon and Drew Bagnell “A reduction of imitation learning and structured prediction to no-regret online learning” In Proceedings of the fourteenth international conference on artificial intelligence and statistics, 2011, pp. 627–635 JMLR WorkshopConference Proceedings
  39. “Evolution strategies as a scalable alternative to reinforcement learning” In arXiv preprint arXiv:1703.03864, 2017
  40. “Natural evolution strategies” In The Journal of Machine Learning Research 15.1 JMLR. org, 2014, pp. 949–980

Summary

  • The paper introduces a decentralized, environment-centric navigation paradigm that employs a GRNN to learn traffic rules, enabling scalable, low-resource multi-agent coordination.
  • It combines imitation and reinforcement learning to dynamically modulate agent velocities, ensuring collision avoidance and reduced congestion.
  • The approach successfully coordinates up to 240 agents, demonstrating its potential for efficient autonomous robotics and smart traffic systems.

Introduction to Neural Traffic Rules

In the domain of robotics, the tasks concerned with guiding multiple agents or robots through environments without collisions are known as multi-agent navigation challenges. This task is crucial across a variety of industries, ranging from automated warehousing systems to the development of autonomous vehicles and the construction of smart cities.

Addressing Scalability for Agents with Limited Resources

When it comes to navigating multiple agents, a central issue is ensuring that each can compute and follow a complex policy individually. Typically, such policies are learned and require each agent to have the computational ability to execute deep neural network inferences. However, given that many robot systems operate with limited computational resources, executing these inferences can be impractical due to high costs and efficiency constraints.

Emulating Real-World Traffic Networks

This paper puts forward a novel environment-centric navigation policy that draws inspiration from the rule-based nature of real-world traffic systems. Unlike most current agent-centric approaches which require heavy computational resources, this work suggests learning predefined traffic rules at an environmental level. By applying a Graph Recurrent Neural Network (GRNN) over a segmented environment into blocks, the method focuses on learning and implementing traffic rules that can be followed by agents with minimal computational capabilities. This echoes the real-world scenario where, for example, drivers follow traffic rules that don't require complex individual decision-making and navigation plans.

Training Environment-Centric Navigation Policies

The proposed approach involves training the GRNN to modulate agents' velocities to ensure collision avoidance and minimize congestion, using a combination of Imitation Learning (IL) and Reinforcement Learning (RL). In cases where groundtruth traffic rules are known, IL can be used to mimic expert behavior. Alternatively, when such expert rules are not available, evolutionary RL is employed to allow the system to seek out and identify effective traffic rules to minimize congestion. Furthermore, the approach demonstrated the ability to coordinate up to 240 agents in simulated environments, proving both scalable and generalizable across different agent and environment configurations.

Conclusion

The key contributions of the paper are threefold: it presents a decentralized navigation paradigm using learned environment-encoded traffic rules, it encapsulates environment-centric navigation policies using GRNN, and it offers a new reward design and training algorithms for these policies within IL and RL settings. The outcome is a scalable and efficient multi-agent navigation method that promises reduced computational demands and negates the necessity for comprehensive inter-agent communication. This paper paves the way for future research directions such as real-world deployment and enhancements to account for time-dependent traffic rules and heterogeneous agent behaviors.