Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
194 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Evolutionary Curriculum Training for DRL-Based Navigation Systems (2306.08870v1)

Published 15 Jun 2023 in cs.RO, cs.AI, cs.LG, and cs.MA

Abstract: In recent years, Deep Reinforcement Learning (DRL) has emerged as a promising method for robot collision avoidance. However, such DRL models often come with limitations, such as adapting effectively to structured environments containing various pedestrians. In order to solve this difficulty, previous research has attempted a few approaches, including training an end-to-end solution by integrating a waypoint planner with DRL and developing a multimodal solution to mitigate the drawbacks of the DRL model. However, these approaches have encountered several issues, including slow training times, scalability challenges, and poor coordination among different models. To address these challenges, this paper introduces a novel approach called evolutionary curriculum training to tackle these challenges. The primary goal of evolutionary curriculum training is to evaluate the collision avoidance model's competency in various scenarios and create curricula to enhance its insufficient skills. The paper introduces an innovative evaluation technique to assess the DRL model's performance in navigating structured maps and avoiding dynamic obstacles. Additionally, an evolutionary training environment generates all the curriculum to improve the DRL model's inadequate skills tested in the previous evaluation. We benchmark the performance of our model across five structured environments to validate the hypothesis that this evolutionary training environment leads to a higher success rate and a lower average number of collisions. Further details and results at our project website.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (30)
  1. Learning and development in neural networks: the importance of starting small. Cognition, 48(1):71–99, 1993. ISSN 0010-0277.
  2. Where to go next: Learning a subgoal recommendation policy for navigation in dynamic environments. IEEE Robotics and Automation Letters, 6(3):4616–4623, 2021.
  3. Crowd-robot interaction: Crowd-aware robot navigation with attention-based deep reinforcement learning. CoRR, abs/1809.08835, 2018.
  4. Decentralized non-communicating multiagent collision avoidance with deep reinforcement learning. CoRR, abs/1609.07845, 2016.
  5. Socially aware motion planning with deep reinforcement learning. CoRR, abs/1703.08862, 2017.
  6. Learning navigation behaviors end to end. CoRR, abs/1809.10124, 2018.
  7. Motion planning among dynamic, decision-making agents with deep reinforcement learning. In 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 3052–3059, 2018.
  8. Collision avoidance in pedestrian-rich environments with deep reinforcement learning. IEEE Access, 9:10357–10377, 2021.
  9. Detection, prediction, and avoidance of dynamic obstacles in urban environments. In 2008 IEEE Intelligent Vehicles Symposium, pages 1149–1154, 2008. doi: 10.1109/IVS.2008.4621214.
  10. Motion planning in dynamic environments using velocity obstacles. The international journal of robotics research, 17(7):760–772, 1998.
  11. Barc: Backward reachability curriculum for robotic reinforcement learning. CoRR, abs/1806.06161, 2018.
  12. Arena-rosnav: Towards deployment of deep-reinforcement-learning-based obstacle avoidance into conventional autonomous navigation systems. In 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 6456–6463. IEEE, 2021.
  13. All-in-one: A drl-based control switch combining state-of-the-art navigation planners. In 2022 International Conference on Robotics and Automation (ICRA), pages 2861–2867, 2022. doi: 10.1109/ICRA46639.2022.9811797.
  14. Waypoint models for instruction-guided navigation in continuous environments. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pages 15162–15171, October 2021.
  15. Holistic deep-reinforcement-learning-based training of autonomous navigation systems, 2023.
  16. Towards optimally decentralized multi-robot collision avoidance via deep reinforcement learning. In 2018 IEEE International Conference on Robotics and Automation (ICRA), page 6252–6259. IEEE Press, 2018.
  17. Evolutionary population curriculum for scaling multi-agent reinforcement learning. CoRR, abs/2003.10423, 2020.
  18. Eitan Marder-Eppstein. Ros move_base package. 2020.
  19. Predictive collision avoidance for the dynamic window approach. pages 8620–8626, 05 2019. doi: 10.1109/ICRA.2019.8794386.
  20. ROS: an open-source Robot Operating System. In Proc. of the IEEE Intl. Conf. on Robotics and Automation (ICRA) Workshop on Open Source Robotics, Kobe, Japan, May 2009.
  21. Adversarial generation of natural language. CoRR, abs/1705.10929, 2017.
  22. Multi-agent motion planning for dense and dynamic environments via deep reinforcement learning. IEEE Robotics and Automation Letters, 5(2):3221–3226, 2020.
  23. Curriculum learning: A survey. CoRR, abs/2101.10382, 2021.
  24. Socially compliant navigation through raw depth inputs with generative adversarial imitation learning. In 2018 IEEE International Conference on Robotics and Automation (ICRA), pages 1111–1117, 2018. doi: 10.1109/ICRA.2018.8460968.
  25. A. T. M. Tsardoulias and C. Zalidis. stdr_simulator - ros_wiki, 2014. URL http://wiki.ros.org/stdr_simulator.
  26. Reciprocal n-body collision avoidance. In Cédric Pradalier, Roland Siegwart, and Gerhard Hirzinger, editors, Robotics Research, pages 3–19, Berlin, Heidelberg, 2011. Springer Berlin Heidelberg.
  27. Paired open-ended trailblazer (POET): endlessly generating increasingly complex and diverse learning environments and their solutions. CoRR, abs/1901.01753, 2019a.
  28. Dynamic curriculum learning for imbalanced data classification. In 2019 IEEE/CVF International Conference on Computer Vision (ICCV), pages 5016–5025, 2019b. doi: 10.1109/ICCV.2019.00512.
  29. Potential gap: A gap-informed reactive policy for safe hierarchical navigation. IEEE Robotics and Automation Letters, 6(4):8325–8332, 2021.
  30. Automatic curriculum learning through value disagreement. CoRR, abs/2006.09641, 2020.
Citations (1)

Summary

We haven't generated a summary for this paper yet.