Papers
Topics
Authors
Recent
2000 character limit reached

NaviSTAR: Socially Aware Robot Navigation with Hybrid Spatio-Temporal Graph Transformer and Preference Learning (2304.05979v2)

Published 12 Apr 2023 in cs.RO

Abstract: Developing robotic technologies for use in human society requires ensuring the safety of robots' navigation behaviors while adhering to pedestrians' expectations and social norms. However, maintaining real-time communication between robots and pedestrians to avoid collisions can be challenging. To address these challenges, we propose a novel socially-aware navigation benchmark called NaviSTAR, which utilizes a hybrid Spatio-Temporal grAph tRansformer (STAR) to understand interactions in human-rich environments fusing potential crowd multi-modal information. We leverage off-policy reinforcement learning algorithm with preference learning to train a policy and a reward function network with supervisor guidance. Additionally, we design a social score function to evaluate the overall performance of social navigation. To compare, we train and test our algorithm and other state-of-the-art methods in both simulator and real-world scenarios independently. Our results show that NaviSTAR outperforms previous methods with outstanding performance\footnote{The source code and experiment videos of this work are available at: https://sites.google.com/view/san-navistar

Definition Search Book Streamline Icon: https://streamlinehq.com
References (27)
  1. A. Garrell and A. Sanfeliu, “Cooperative social robots to accompany groups of people,” The International Journal of Robotics Research, vol. 31, no. 13, pp. 1675–1701, 2012.
  2. N. E. Du Toit and J. W. Burdick, “Robot motion planning in dynamic, uncertain environments,” IEEE Transactions on Robotics, vol. 28, no. 1, pp. 101–115, 2011.
  3. M. Bennewitz, W. Burgard, G. Cielniak, and S. Thrun, “Learning motion patterns of people for compliant robot motion,” The International Journal of Robotics Research, vol. 24, no. 1, pp. 31–48, 2005.
  4. P. Trautman, J. Ma, R. M. Murray, and A. Krause, “Robot navigation in dense human crowds: Statistical models and experimental studies of human–robot cooperation,” The International Journal of Robotics Research, vol. 34, no. 3, pp. 335–356, 2015.
  5. H. Kretzschmar, M. Spies, C. Sprunk, and W. Burgard, “Socially compliant mobile robot navigation via inverse reinforcement learning,” The International Journal of Robotics Research, vol. 35, no. 11, pp. 1289–1307, 2016.
  6. Y. F. Chen, M. Liu, M. Everett, and J. P. How, “Decentralized non-communicating multiagent collision avoidance with deep reinforcement learning,” in 2017 IEEE international conference on robotics and automation (ICRA).   IEEE, 2017, pp. 285–292.
  7. C. Chen, Y. Liu, S. Kreiss, and A. Alahi, “Crowd-robot interaction: Crowd-aware robot navigation with attention-based deep reinforcement learning,” in 2019 International Conference on Robotics and Automation (ICRA).   IEEE, 2019, pp. 6015–6022.
  8. C. Chen, S. Hu, P. Nikdel, G. Mori, and M. Savva, “Relational graph learning for crowd navigation,” in 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).   IEEE, 2020.
  9. S. Liu, P. Chang, W. Liang, N. Chakraborty, and K. Driggs-Campbell, “Decentralized structural-rnn for robot crowd navigation with deep reinforcement learning,” in 2021 IEEE International Conference on Robotics and Automation (ICRA).   IEEE, 2021, pp. 3517–3524.
  10. M. Sun, F. Baldini, P. Trautman, and T. Murphey, “Move Beyond Trajectories: Distribution Space Coupling for Crowd Navigation,” in Proceedings of Robotics: Science and Systems, Virtual, July 2021.
  11. R. Wang, W. Wang, and B.-C. Min, “Feedback-efficient active preference learning for socially aware robot navigation,” in 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).   IEEE, 2022, pp. 11 336–11 343.
  12. S. Liu, P. Chang, Z. Huang, N. Chakraborty, W. Liang, J. Geng, and K. Driggs-Campbell, “Socially aware robot crowd navigation with interaction graphs and human trajectory prediction,” arXiv preprint arXiv:2203.01821, 2022.
  13. C. Yu, X. Ma, J. Ren, H. Zhao, and S. Yi, “Spatio-temporal graph transformer networks for pedestrian trajectory prediction,” in European Conference on Computer Vision.   Springer, 2020, pp. 507–523.
  14. Z. Li, W. Wang, H. Li, E. Xie, C. Sima, T. Lu, Y. Qiao, and J. Dai, “Bevformer: Learning bird’s-eye-view representation from multi-camera images via spatiotemporal transformers,” in Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23–27, 2022.   Springer, 2022, pp. 1–18.
  15. C. Chen, Y. Liu, L. Chen, and C. Zhang, “Bidirectional spatial-temporal adaptive transformer for urban traffic flow forecasting,” IEEE Transactions on Neural Networks and Learning Systems, 2022.
  16. Y.-H. H. Tsai, S. Bai, P. P. Liang, J. Z. Kolter, L.-P. Morency, and R. Salakhutdinov, “Multimodal transformer for unaligned multimodal language sequences,” in Proceedings of the conference. Association for Computational Linguistics. Meeting, vol. 2019.   NIH Public Access, 2019, p. 6558.
  17. R. J. Chen, M. Y. Lu, W.-H. Weng, T. Y. Chen, D. F. Williamson, T. Manz, M. Shady, and F. Mahmood, “Multimodal co-attention transformer for survival prediction in gigapixel whole slide images,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 4015–4025.
  18. R. Wang, W. Jo, D. Zhao, W. Wang, B. Yang, G. Chen, and B.-C. Min, “Husformer: A multi-modal transformer for multi-modal human state recognition,” arXiv preprint arXiv:2209.15182, 2022.
  19. A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin, “Attention is all you need,” Advances in neural information processing systems, vol. 30, 2017.
  20. T. N. Kipf and M. Welling, “Semi-supervised classification with graph convolutional networks,” in International Conference on Learning Representations, 2017.
  21. K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778.
  22. K. Lee, L. Smith, A. Dragan, and P. Abbeel, “B-pref: Benchmarking preference-based reinforcement learning,” in Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 1), 2021.
  23. T. Haarnoja, A. Zhou, P. Abbeel, and S. Levine, “Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor,” in International conference on machine learning.   PMLR, 2018, pp. 1861–1870.
  24. J. Van Den Berg, S. J. Guy, M. Lin, and D. Manocha, “Reciprocal n-body collision avoidance,” in Robotics Research: The 14th International Symposium ISRR.   Springer, 2011, pp. 3–19.
  25. J. Rios-Martinez, A. Spalanzani, and C. Laugier, “From proxemics theory to socially-aware navigation: A survey,” International Journal of Social Robotics, vol. 7, pp. 137–153, 2015.
  26. A. Pramanik, S. K. Pal, J. Maiti, and P. Mitra, “Granulated rcnn and multi-class deep sort for multi-object detection and tracking,” IEEE Transactions on Emerging Topics in Computational Intelligence, vol. 6, no. 1, pp. 171–181, 2021.
  27. L. Bertoni, S. Kreiss, T. Mordan, and A. Alahi, “Monstereo: When monocular and stereo meet at the tail of 3d human localization,” in 2021 IEEE International Conference on Robotics and Automation (ICRA).   IEEE, 2021, pp. 5126–5132.
Citations (10)

Summary

  • The paper demonstrates a novel hybrid spatio-temporal graph transformer combined with preference learning to enhance socially aware robot navigation.
  • It integrates spatial and temporal features via a fully connected graph to capture and model human-robot interactions, boosting navigation success and social compliance.
  • Empirical tests show NaviSTAR outperforms existing methods, achieving higher success rates and social scores in both open and constrained environments.

Overview of NaviSTAR: Socially Aware Robot Navigation with Hybrid Spatio-Temporal Graph Transformer and Preference Learning

This essay examines the research presented in the paper titled "NaviSTAR: Socially Aware Robot Navigation with Hybrid Spatio-Temporal Graph Transformer and Preference Learning." The paper introduces a novel benchmark called NaviSTAR that addresses the complex challenges of socially aware robot navigation by utilizing a hybrid Spatio-Temporal Graph Transformer and preference learning. The primary focus lies in ensuring that robots navigate safely within human environments while adhering to social norms and pedestrian expectations.

The proposed approach leverages the Spatio-Temporal Graph Transformer to capture intricate interactions between humans and robots, thereby comprehensively modeling human-robot interaction (HRI). This framework is augmented by a preference learning mechanism that effectively encodes human expectations and social norms into the decision-making process of robotic navigation systems.

Methodology

The authors distinguish their work from existing methodologies by highlighting the ability of NaviSTAR to integrate spatial and temporal features through a fully connected graph representation of HRI. The core technological innovation lies in the use of a Spatio-Temporal Graph Transformer combined with a multi-modal transformer. This amalgamation allows for the capture of long-term dependencies and fusion of heterogeneous spatial and temporal features inherent to dynamic, human-filled environments.

Numerical results presented in the paper indicate that NaviSTAR exhibits superior performance compared to existing state-of-the-art methods in both simulated and real-world environments. Specifically, the results highlight improvements in both navigation success rates and social compliance as assessed by a novel social score function.

Numerical Results and Claims

The authors present concrete numerical evidence supporting the efficacy of NaviSTAR. In a series of 500 tests conducted under varying conditions, the algorithm demonstrated a higher success rate and improved social scores relative to traditional methods such as CADRL and SARL, as well as more recent approaches like SRNN. Notably, NaviSTAR outperformed these baseline methods not only in open spaces but also in constrained environments with varied fields of view.

The proposed model's superiority is attributed to the sophisticated representation of agent interactions via the Spatio-Temporal Graph Transformer network. This was evident in the visualization of spatial-temporal and cross-modal attention matrices, which showed how the system could accurately interpret and predict interactions and dependencies within a crowd, a capability that is crucial for ensuring safe and socially acceptable navigation.

Additionally, the inclusion of preference learning into the reinforcement learning setup is posited to yield a more natural and desirable robotic behavior by adjusting the reward function based on human feedback.

Implications and Future Developments

The implications of this research are significant for the field of socially aware robot navigation. The ability to seamlessly integrate into human-rich environments while respecting social norms is a pivotal requirement for the deployment of service robots in public spaces. The methodology introduced by NaviSTAR can potentially be extended to other domains requiring complex interaction modeling and decision-making, such as autonomous vehicles and assistive robotics.

Future research could focus on enhancing the scalability of NaviSTAR to handle even more complex environments with a larger number of interacting agents. Additionally, investigating the integration of other forms of human feedback and adaptive learning could further refine the system's ability to comply with diverse social expectations.

In summary, the NaviSTAR framework represents a significant advancement in socially aware navigation by adeptly combining Spatio-Temporal Graph Transformers with preference learning. The empirical evidence provided underscores its capability to outperform existing models in achieving both efficient and socially compliant navigation, marking an important step forward in autonomous robotic systems' ability to interact harmoniously with humans.

Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Youtube Logo Streamline Icon: https://streamlinehq.com