Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
139 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Autonomous Port Navigation With Ranging Sensors Using Model-Based Reinforcement Learning (2312.05257v1)

Published 17 Nov 2023 in cs.RO and cs.AI

Abstract: Autonomous shipping has recently gained much interest in the research community. However, little research focuses on inland - and port navigation, even though this is identified by countries such as Belgium and the Netherlands as an essential step towards a sustainable future. These environments pose unique challenges, since they can contain dynamic obstacles that do not broadcast their location, such as small vessels, kayaks or buoys. Therefore, this research proposes a navigational algorithm which can navigate an inland vessel in a wide variety of complex port scenarios using ranging sensors to observe the environment. The proposed methodology is based on a machine learning approach that has recently set benchmark results in various domains: model-based reinforcement learning. By randomizing the port environments during training, the trained model can navigate in scenarios that it never encountered during training. Furthermore, results show that our approach outperforms the commonly used dynamic window approach and a benchmark model-free reinforcement learning algorithm. This work is therefore a significant step towards vessels that can navigate autonomously in complex port scenarios.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (30)
  1. Assessment of potential of maritime and inland ports and inland waterways and of related policy measures, including industrial policy measures : final report. Publications Office (2020). URL https://data.europa.eu/doi/10.2832/03796.
  2. ‘‘Towards the assessment of potential impact of unmanned vessels on maritime transportation safety.’’ Reliability Engineering & System Safety Vol. 165 (2017): pp. 155–169. 10.1016/j.ress.2017.03.029.
  3. ‘‘Costs and benefits of autonomous shipping—a literature review.’’ Applied Sciences Vol. 11 No. 10 (2021): p. 4553. 10.3390/app11104553.
  4. ‘‘Deep learning for autonomous ship-oriented small ship detection.’’ Safety Science Vol. 130 (2020): p. 104812. 10.1016/j.ssci.2020.104812.
  5. ‘‘Deep reinforcement learning-based collision avoidance for an autonomous ship.’’ Ocean Engineering Vol. 234 (2021): p. 109216. 10.1016/j.oceaneng.2021.109216.
  6. ‘‘A human-like collision avoidance method for autonomous ship with attention-based deep reinforcement learning.’’ Ocean Engineering Vol. 264 (2022): p. 112378.
  7. Reinforcement learning: An introduction. MIT press (2018).
  8. ‘‘Deep Reinforcement Learning with Double Q-Learning.’’ Proceedings of the AAAI Conference on Artificial Intelligence Vol. 30 No. 1. 10.1609/aaai.v30i1.10295.
  9. ‘‘Proximal policy optimization algorithms.’’ arXiv preprint arXiv:1707.06347 10.48550/arXiv.1707.06347.
  10. ‘‘Model-based reinforcement learning: A survey.’’ arXiv preprint arXiv:2006.16712 10.48550/arXiv.2006.16712.
  11. ‘‘Mastering atari, go, chess and shogi by planning with a learned model.’’ Nature Vol. 588 No. 7839 (2020): pp. 604–609. 10.1038/s41586-020-03051-4.
  12. ‘‘Model-based reinforcement learning via meta-policy optimization.’’ Conference on Robot Learning: pp. 617–629. 2018. PMLR. URL https://proceedings.mlr.press/v87/clavera18a.html.
  13. ‘‘Collision Avoidance Algorithm Based on COLREGs for Unmanned Surface Vehicle.’’ Journal of Marine Science and Engineering Vol. 9 No. 8. 10.3390/jmse9080863.
  14. ‘‘A novel path following approach for autonomous ships based on fast marching method and deep reinforcement learning.’’ Ocean Engineering Vol. 257 (2022): p. 111495. 10.1016/j.oceaneng.2022.111495.
  15. Sethian, James A. ‘‘A fast marching level set method for monotonically advancing fronts.’’ Proceedings of the National Academy of Sciences Vol. 93 No. 4 (1996): pp. 1591–1595. 10.1073/pnas.93.4.1591.
  16. ‘‘Line-of-sight path following of underactuated marine craft.’’ IFAC Proceedings Volumes Vol. 36 No. 21 (2003): pp. 211–216. 10.1016/S1474-6670(17)37809-6. 6th IFAC Conference on Manoeuvring and Control of Marine Craft (MCMC 2003), Girona, Spain, 17-19 September, 1997.
  17. ‘‘Model-based Deep Reinforcement Learning for Data-driven Motion Control of an Under-actuated Unmanned Surface Vehicle: Path Following and Trajectory Tracking.’’ Journal of the Franklin Institute 10.1016/j.jfranklin.2022.10.020.
  18. ‘‘Nonlinear Model Predictive Control for trajectory tracking and collision avoidance of underactuated vessels with disturbances.’’ Ocean Engineering Vol. 160 (2018): pp. 168–180. 10.1016/j.oceaneng.2018.04.026.
  19. ‘‘A human-like collision avoidance method for autonomous ship with attention-based deep reinforcement learning.’’ Ocean Engineering Vol. 264 (2022): p. 112378. 10.1016/j.oceaneng.2022.112378.
  20. ‘‘Safety Aware Autonomous Path Planning Using Model Predictive Reinforcement Learning for Inland Waterways.’’ IECON 2022 – 48th Annual Conference of the IEEE Industrial Electronics Society: pp. 1–6. 2022. 10.1109/IECON49645.2022.9968678.
  21. ‘‘Decision-Making for the Autonomous Navigation of Maritime Autonomous Surface Ships Based on Scene Division and Deep Reinforcement Learning.’’ Sensors Vol. 19 No. 18. 10.3390/s19184055.
  22. ‘‘Sampling-based algorithms for optimal motion planning.’’ The international journal of robotics research Vol. 30 No. 7 (2011): pp. 846–894. 10.1177/0278364911406761.
  23. ‘‘Nonlinear model predictive control.’’ Nonlinear model predictive control. Springer (2017): pp. 45–69. 10.1007/978-3-642-01094-1.
  24. ‘‘An overview of moos-ivp and a brief users guide to the ivp helm autonomy software.’’ .
  25. ‘‘Domain randomization for transferring deep neural networks from simulation to the real world.’’ 2017 IEEE/RSJ international conference on intelligent robots and systems (IROS): pp. 23–30. 2017. IEEE. 10.1109/IROS.2017.8202133.
  26. ‘‘Monte-carlo tree search: A new framework for game ai.’’ Proceedings of the AAAI Conference on Artificial Intelligence and Interactive Digital Entertainment, Vol. 4.  1: pp. 216–217. 2008. 10.1609/aiide.v4i1.18700.
  27. ‘‘Monte-Carlo tree search as regularized policy optimization.’’ International Conference on Machine Learning: pp. 3769–3778. 2020. PMLR. URL https://proceedings.mlr.press/v119/grill20a.html.
  28. ‘‘Adam: A method for stochastic optimization.’’ arXiv preprint arXiv:1412.6980 10.48550/arXiv.1412.6980.
  29. ‘‘Learning internal representations by error propagation.’’ Technical report no. California Univ San Diego La Jolla Inst for Cognitive Science. 1985. URL https://ieeexplore.ieee.org/document/6302929.
  30. ‘‘A Hybrid Path Planning Algorithm for Unmanned Surface Vehicles in Complex Environment With Dynamic Obstacles.’’ IEEE Access Vol. 7 (2019): pp. 126439–126449. 10.1109/ACCESS.2019.2936689.
Citations (1)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com