Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
126 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Improving Generalization in Reinforcement Learning Training Regimes for Social Robot Navigation (2308.14947v2)

Published 29 Aug 2023 in cs.RO, cs.LG, and cs.MA

Abstract: In order for autonomous mobile robots to navigate in human spaces, they must abide by our social norms. Reinforcement learning (RL) has emerged as an effective method to train sequential decision-making policies that are able to respect these norms. However, a large portion of existing work in the field conducts both RL training and testing in simplistic environments. This limits the generalization potential of these models to unseen environments, and the meaningfulness of their reported results. We propose a method to improve the generalization performance of RL social navigation methods using curriculum learning. By employing multiple environment types and by modeling pedestrians using multiple dynamics models, we are able to progressively diversify and escalate difficulty in training. Our results show that the use of curriculum learning in training can be used to achieve better generalization performance than previous training methods. We also show that results presented in many existing state-of-the-art RL social navigation works do not evaluate their methods outside of their training environments, and thus do not reflect their policies' failure to adequately generalize to out-of-distribution scenarios. In response, we validate our training approach on larger and more crowded testing environments than those used in training, allowing for more meaningful measurements of model performance.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (36)
  1. Social momentum: Design and evaluation of a framework for socially competent robot navigation. J. Hum.-Robot Interact., 11(2), feb 2022.
  2. Social force model for pedestrian dynamics. Physical Review E, 51(5):4282–4286, May 1995.
  3. Reciprocal n-body collision avoidance. In Robotics research, pages 3–19. Springer, 2011.
  4. Crowd-robot interaction: Crowd-aware robot navigation with attention-based deep reinforcement learning. In 2019 International Conference on Robotics and Automation (ICRA), pages 6015–6022. IEEE, 2019.
  5. Decentralized non-communicating multiagent collision avoidance with deep reinforcement learning. In 2017 IEEE international conference on robotics and automation (ICRA), pages 285–292. IEEE, 2017.
  6. Motion planning among dynamic, decision-making agents with deep reinforcement learning. In 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 3052–3059. IEEE, 2018.
  7. Core challenges of social robot navigation: A survey. CoRR, abs/2103.05668, 2021.
  8. J. Borenstein and Y. Koren. Real-time obstacle avoidance for fast mobile robots. IEEE Transactions on Systems, Man, and Cybernetics, 19(5):1179–1187, 1989.
  9. The dynamic window approach to collision avoidance. IEEE Robotics Automation Magazine, 4(1):23–33, 1997.
  10. Probabilistic algorithms and the interactive museum tour-guide robot minerva. The International Journal of Robotics Research, 19(11):972–999, 2000.
  11. The museum tour-guide robot rhino. In Autonome Mobile Systeme 1998, pages 245–254. Springer, 1999.
  12. Franck Feurtey. Simulating the collision avoidance behavior of pedestrians. Master’s thesis, University of Tokyo, Department of Electronic Engineering, 2000.
  13. Social lstm: Human trajectory prediction in crowded spaces. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 961–971, 2016.
  14. Social attention: Modeling attention in human crowds. In 2018 IEEE international Conference on Robotics and Automation (ICRA), pages 4601–4607. IEEE, 2018.
  15. A probabilistic model of human motion and navigation intent for mobile robot path planning. In 2009 4th International Conference on Autonomous Robots and Agents, pages 663–668. IEEE, 2009.
  16. Learning motion patterns of people for compliant robot motion. The International Journal of Robotics Research, 24(1):31–48, 2005.
  17. Human-robot co-navigation using anticipatory indicators of human walking motion. In 2015 IEEE International Conference on Robotics and Automation (ICRA), pages 6183–6190. IEEE, 2015.
  18. Unfreezing the robot: Navigation in dense, interacting crowds. In 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems, pages 797–803. IEEE, 2010.
  19. Socially compliant mobile robot navigation via inverse reinforcement learning. The International Journal of Robotics Research, 35(11):1289–1307, 2016.
  20. Socially adaptive path planning in human environments using inverse reinforcement learning. International Journal of Social Robotics, 8(1):51–66, 2016.
  21. Planning-based prediction for pedestrians. In 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems, pages 3931–3936. IEEE, 2009.
  22. Socially compliant navigation through raw depth inputs with generative adversarial imitation learning. In 2018 IEEE International Conference on Robotics and Automation (ICRA), pages 1111–1117. IEEE, 2018.
  23. Deep-learned collision avoidance policy for distributed multiagent navigation. IEEE Robotics and Automation Letters, 2(2):656–663, 2017.
  24. Map-based deep imitation learning for obstacle avoidance. In 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 8644–8649. IEEE, 2018.
  25. Maximum entropy inverse reinforcement learning. In Aaai, volume 8, pages 1433–1438. Chicago, IL, USA, 2008.
  26. Risk-sensitive sequential action control with multi-modal human trajectory forecasting for safe crowd-robot interaction. In 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 11205–11212. IEEE, 2020.
  27. Real-time navigation of independent agents using adaptive roadmaps. In ACM SIGGRAPH 2008 classes, pages 99–106, 11 2007.
  28. Robot companion: A social-force based approach with human awareness-navigation in crowded environments. In 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, pages 1688–1694. IEEE, 2013.
  29. Robot social-aware navigation framework to accompany people walking side-by-side. Autonomous robots, 41(4):775–793, 2017.
  30. Socially aware motion planning with deep reinforcement learning. In 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 1343–1350. IEEE, 2017.
  31. Reciprocal velocity obstacles for real-time multi-agent navigation. In 2008 IEEE International Conference on Robotics and Automation, pages 1928–1935, 2008.
  32. Sven Kreiss. Deep social force. CoRR, abs/2109.12081, 2021.
  33. Walking ahead: The headed social force model. PloS one, 12(1):e0169734, 2017.
  34. E.T. Hall and Society for the Anthropology of Visual Communication. Handbook for Proxemic Research. Society for Visual Anthropology Series. Society for the Anthropology of Visual Communication, 1974.
  35. Proxemics models for human-aware navigation in robotics: Grounding interaction and personal space models in experimental data from psychology. In Proceedings of the 3rd IROS’2014 workshop “Assistance and Service Robotics in a Human Environment”, Chicago, United States, September 2014.
  36. Sesno: Sample efficient social navigation from observation. In 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 9164–9171, 2022.
Citations (1)

Summary

We haven't generated a summary for this paper yet.