A Transferability Metric Using Scene Similarity and Local Map Observation for DRL Navigation (2306.04910v4)
Abstract: While deep reinforcement learning (DRL) has attracted a rapidly growing interest in solving the problem of navigation without global maps, DRL typically leads to a mediocre navigation performance in practice due to the gap between the training scene and the actual test scene. To quantify the transferability of a DRL agent between the training and test scenes, this paper proposes a new transferability metric -- the scene similarity calculated using an improved image template matching algorithm. Specifically, two transferability performance indicators are designed including the global scene similarity that evaluates the overall robustness of a DRL algorithm and the local scene similarity that serves as a safety measure when a DRL agent is deployed without a global map. In addition, this paper proposes the use of a local map that fuses 2D LiDAR data with spatial information of both the agent and the destination as the DRL observation, aiming to improve the transferability of DRL navigation algorithms. With a wheeled robot as the case study platform, both simulation and real-world experiments are conducted in a total of 26 different scenes. The experimental results affirm the robustness of the local map observation design and demonstrate the strong correlation between the scene similarity metric and the success rate of DRL navigation algorithms.
- Y. Chen, G. Chen, L. Pan, J. Ma, Y. Zhang, Y. Zhang, and J. Ji, “DRQN-based 3D obstacle avoidance with a limited field of view,” in 2021 IEEE/RSJ Int. Conf. Intell. Robots Syst. (IROS), pp. 8137–8143.
- D. Hoeller, L. Wellhausen, F. Farshidian, and M. Hutter, “Learning a state representation and navigation in cluttered and dynamic environments,” IEEE Robot. Automat. Lett., vol. 6, no. 3, pp. 5081–5088, 2021.
- F. Leiva and J. Ruiz-del Solar, “Robust RL-based map-less local planning: Using 2d point clouds as observations,” IEEE Robot. Automat. Lett., vol. 5, no. 4, pp. 5787–5794, 2020.
- J. Choi, K. Park, M. Kim, and S. Seok, “Deep reinforcement learning of navigation in a complex and crowded environment with a limited field of view,” in 2019 Int. Conf. Robot. Automat. (ICRA), pp. 5993–6000.
- L. Tai, G. Paolo, and M. Liu, “Virtual-to-real deep reinforcement learning: Continuous control of mobile robots for mapless navigation,” in 2017 IEEE/RSJ Int. Conf. Intell. Robots Syst. (IROS), pp. 31–36.
- M. Pfeiffer, S. Shukla, M. Turchetta, C. Cadena, A. Krause, R. Siegwart, and J. Nieto, “Reinforced imitation: Sample efficient deep reinforcement learning for mapless navigation by leveraging prior demonstrations,” IEEE Robot. Automat. Lett., vol. 3, no. 4, pp. 4423–4430, 2018.
- A. Wahid, A. Toshev, M. Fiser, and T.-W. E. Lee, “Long range neural navigation policies for the real world,” in 2019 IEEE/RSJ Int. Conf. Intell. Robots Syst. (IROS), pp. 82–89.
- Y. Jang, J. Baek, and S. Han, “Hindsight intermediate targets for mapless navigation with deep reinforcement learning,” IEEE Trans. Ind. Electron., vol. 69, no. 11, pp. 11 816–11 825, 2022.
- K. Lobos-Tsunekawa and T. Harada, “Point cloud based reinforcement learning for sim-to-real and partial observability in visual navigation,” in 2020 IEEE/RSJ Int. Conf. Intell. Robots Syst. (IROS), pp. 5871–5878.
- W. Zhang, Y. Zhang, N. Liu, K. Ren, and P. Wang, “IPAPRec: A promising tool for learning high-performance mapless navigation skills with deep reinforcement learning,” IEEE/ASME Trans. Mechatron., vol. 27, no. 6, pp. 5451–5461, 2022.
- K. Zhu and T. Zhang, “Deep reinforcement learning based mobile robot navigation: A review,” Tsinghua Sci. Technol., vol. 26, no. 5, pp. 674–691, 2021.
- C. Warren, “Global path planning using artificial potential fields,” in 1989 Int. Conf. Robot. and Automat., vol. 1, pp. 316–321.
- D. Fox, W. Burgard, and S. Thrun, “The dynamic window approach to collision avoidance,” IEEE Robot. Automat. Mag., vol. 4, no. 1, pp. 23–33, 1997.
- Y. Zhu, Z. Wang, C. Chen, and D. Dong, “Rule-based reinforcement learning for efficient robot navigation with space reduction,” IEEE/ASME Trans. Mechatron., vol. 27, no. 2, pp. 846–857, 2022.
- H. Jiang, M. A. Esfahani, K. Wu, K. wah Wan, K. kian Heng, H. Wang, and X. Jiang, “iTD3-CLN: Learn to navigate in dynamic scene through deep reinforcement learning,” Neurocomputing, vol. 503, pp. 118–128, 2022.
- D. Liu, Z. Lyu, Q. Zou, X. Bian, M. Cong, and Y. Du, “Robotic navigation based on experiences and predictive map inspired by spatial cognition,” IEEE/ASME Trans. Mechatron., vol. 27, no. 6, pp. 4316–4326, 2022.
- M. Pfeiffer, G. Paolo, H. Sommer, J. Nieto, R. Siegwart, and C. Cadena, “A data-driven model for interaction-aware pedestrian motion prediction in object cluttered environments,” in 2018 Int. Conf. Robot. Automat. (ICRA), pp. 5921–5928.
- S. Yao, G. Chen, Q. Qiu, J. Ma, X. Chen, and J. Ji, “Crowd-aware robot navigation for pedestrians with multiple collision avoidance strategies via map-based deep reinforcement learning,” in 2021 IEEE/RSJ Int. Conf. Intell. Robots Syst. (IROS), pp. 8144–8150.
- Y. Chebotar, A. Handa, V. Makoviychuk, M. Macklin, J. Issac, N. Ratliff, and D. Fox, “Closing the sim-to-real loop: Adapting simulation randomization with real world experience,” in 2019 Int. Conf. Robot. Automat. (ICRA), pp. 8973–8979.
- B. Balaji, S. Mallya, S. Genc, S. Gupta, L. Dirac, V. Khare, G. Roy, T. Sun, Y. Tao, B. Townsend, E. Calleja, S. Muralidhara, and D. Karuppasamy, “Deepracer: Autonomous racing platform for experimentation with sim2real reinforcement learning,” in 2020 IEEE Int. Conf. Robot. Automat. (ICRA), pp. 2746–2754.
- Z. Ding, N. F. Lepora, and E. Johns, “Sim-to-real transfer for optical tactile sensing,” in 2020 IEEE Int. Conf. Robot. Automat. (ICRA), pp. 1639–1645.
- T. Chaffre, J. Moras, A. Chan-Hon-Tong, and J. Marzat, “Sim-to-real transfer with incremental environment complexity for reinforcement learning of depth-based robot navigation,” 2020, arXiv:abs/2004.14684.
- N. Q. Hieu, D. T. Hoang, D. Niyato, P. Wang, D. I. Kim, and C. Yuen, “Transferable deep reinforcement learning framework for autonomous vehicles with joint radar-data communications,” IEEE Trans. Commun., vol. 70, no. 8, pp. 5164–5180, 2022.
- J. Zhang, L. Tai, P. Yun, Y. Xiong, M. Liu, J. Boedecker, and W. Burgard, “Vr-goggles for robots: Real-to-sim domain adaptation for visual control,” IEEE Robot. Automat. Lett., vol. 4, no. 2, pp. 1148–1155, 2019.
- Y. Park, S. H. Lee, and I. H. Suh, “Sim-to-real visual grasping via state representation learning based on combining pixel-level and feature-level domain adaptation,” in 2021 IEEE Int. Conf. Robot. Automat. (ICRA), pp. 6300–6307.
- H. Bharadhwaj, Z. Wang, Y. Bengio, and L. Paull, “A data-efficient framework for training and sim-to-real transfer of navigation policies,” in 2019 Int. Conf. Robot. Automat. (ICRA), pp. 782–788.
- Q. Luo, M. Sorokin, and S. Ha, “A few shot adaptation of visual navigation skills to new observations using meta-learning,” in 2021 IEEE Int. Conf. Robot. Automat. (ICRA), pp. 13 231–13 237.
- V. Mnih, K. Kavukcuoglu, D. Silver, A. Graves, I. Antonoglou, D. Wierstra, and M. A. Riedmiller, “Playing atari with deep reinforcement learning,” 2013, arXiv:abs/1312.5602.
- S. Fujimoto, H. van Hoof, and D. Meger, “Addressing function approximation error in actor-critic methods,” in Proc. 35th Int. Conf. Mach. Learn., J. Dy and A. Krause, Eds., vol. 80. PMLR, 10–15 Jul 2018, pp. 1587–1596.
- G. Grisetti, C. Stachniss, and W. Burgard, “Improved techniques for grid mapping with rao-blackwellized particle filters,” IEEE Trans. Robot., vol. 23, no. 1, pp. 34–46, 2007.
- Shiwei Lian (3 papers)
- Feitian Zhang (16 papers)