Autonomous robotic re-alignment for face-to-face underwater human-robot interaction (2401.04320v1)
Abstract: The use of autonomous underwater vehicles (AUVs) to accomplish traditionally challenging and dangerous tasks has proliferated thanks to advances in sensing, navigation, manipulation, and on-board computing technologies. Utilizing AUVs in underwater human-robot interaction (UHRI) has witnessed comparatively smaller levels of growth due to limitations in bi-directional communication and significant technical hurdles to bridge the gap between analogies with terrestrial interaction strategies and those that are possible in the underwater domain. A necessary component to support UHRI is establishing a system for safe robotic-diver approach to establish face-to-face communication that considers non-standard human body pose. In this work, we introduce a stereo vision system for enhancing UHRI that utilizes three-dimensional reconstruction from stereo image pairs and machine learning for localizing human joint estimates. We then establish a convention for a coordinate system that encodes the direction the human is facing with respect to the camera coordinate frame. This allows automatic setpoint computation that preserves human body scale and can be used as input to an image-based visual servo control scheme. We show that our setpoint computations tend to agree both quantitatively and qualitatively with experimental setpoint baselines. The methodology introduced shows promise for enhancing UHRI by improving robotic perception of human orientation underwater.
- M. Matarić, “On relevance: Balancing theory and practice in HRI,” ACM Transactions on Human-Robot Interaction (THRI), vol. 7, no. 1, p. 1–2, 2018.
- A. Cherubini, R. Passama, A. Crosnier, A. Lasnier, and P. Fraisse, “Collaborative manufacturing with physical human–robot interaction,” Robotics and Computer-Integrated Manufacturing, vol. 40, p. 1–13, 2016.
- H. Su, J. Sandoval, M. Makhdoomi, G. Ferrigno, and E. De Momi, “Safety-enhanced human-robot interaction control of redundant robot for teleoperated minimally invasive surgery,” in 2018 IEEE International Conference on Robotics and Automation (ICRA), p. 6611–6616, IEEE, 2018.
- J. Fasola and M. J. Matarić, “A socially assistive robot exercise coach for the elderly,” Journal of Human-Robot Interaction, vol. 2, no. 2, p. 3–32, 2013.
- F. Jentsch, Human-robot interactions in future military operations. CRC Press, 2016.
- M. Fulton, C. Edge, and J. Sattar, “Robot communication via motion: Closing the underwater human-robot interaction loop,” in 2019 International Conference on Robotics and Automation (ICRA), p. 4660–4666, IEEE, 2019.
- L. Takayama, W. Ju, and C. Nass, “Beyond dirty, dangerous and dull: what everyday people think robots should do,” in 2008 3rd ACM/IEEE international conference on human-robot interaction (HRI), p. 25–32, IEEE, 2008.
- T. B. Sheridan, “Human–robot interaction: status and challenges,” Human factors, vol. 58, no. 4, p. 525–532, 2016.
- P. Corke, C. Detweiler, M. Dunbabin, M. Hamilton, D. Rus, and I. Vasilescu, “Experiments with underwater robot localization and tracking,” in Proceedings 2007 IEEE International Conference on Robotics and Automation, p. 4556–4561, IEEE, 2007.
- C. Edge, M. J. Islam, C. Morse, and J. Sattar, “A generative approach for detection-driven underwater image enhancement,” arXiv preprint arXiv:2012.05990, 2020.
- M. J. Islam, Y. Xia, and J. Sattar, “Fast underwater image enhancement for improved visual perception,” IEEE Robotics and Automation Letters, vol. 5, no. 2, pp. 3227–3234, 2020.
- M. J. Islam, P. Luo, and J. Sattar, “Simultaneous Enhancement and Super-Resolution of Underwater Imagery for Improved Visual Perception,” in Proceedings of Robotics: Science and Systems, (Corvalis, Oregon, USA), July 2020.
- M. C. Hastings, “Coming to terms with the effects of ocean noise on marine animals,” Acoustics today, vol. 4, no. 2, p. 22–34, 2008.
- M. Fulton, A. Prabhu, and J. Sattar, “HREyes: Design, Development, and Evaluation of a Novel Method for AUVs to Communicate Information and Gaze Direction,” arXiv preprint arXiv:2211.02946, 2022.
- S. S. Enan, M. Fulton, and J. Sattar, “Robotic Detection of a Human-Comprehensible Gestural Language for Underwater Multi-Human-Robot Collaboration,” arXiv preprint arXiv:2207.05331, 2022.
- M. Fulton, J. Hong, and J. Sattar, “Using Monocular Vision and Human Body Priors for AUVs to Autonomously Approach Divers,” in 2022 International Conference on Robotics and Automation (ICRA), p. 1076–1082, IEEE, 2022.
- S. S. Enan and J. Sattar, “Visual Detection of Diver Attentiveness for Underwater Human-Robot Interaction,” arXiv preprint arXiv:2209.14447, 2022.
- Y. Xia and J. Sattar, “Visual diver recognition for underwater human-robot collaboration,” in 2019 international conference on robotics and automation (ICRA), p. 6839–6845, IEEE, 2019.
- M. J. Islam and J. Sattar, “Mixed-domain biological motion tracking for underwater human-robot interaction,” in 2017 IEEE international conference on robotics and automation (ICRA), p. 4457–4464, IEEE, 2017.
- J. Sattar and G. Dudek, “Where is your dive buddy: tracking humans underwater using spatio-temporal features,” in IEEE/RSJ International Conference on Intelligent Robots and Systems, p. 3654–3659, IEEE, 2007.
- I. Kvasić, N. Mišković, and Z. Vukić, “Convolutional neural network architectures for sonar-based diver detection and tracking,” in OCEANS 2019-Marseille, p. 1–6, IEEE, 2019.
- K. W. Lo and B. G. Ferguson, “Diver detection and localization using passive sonar,” in Proceedings of Acoustics, vol. 8, p. 1–8, 2012.
- K. J. DeMarco, M. E. West, and A. M. Howard, “Sonar-based detection and tracking of a diver for underwater human-robot interaction scenarios,” in 2013 IEEE International Conference on Systems, Man, and Cybernetics, p. 2378–2383, IEEE, 2013.
- M. Fulton, C. Edge, and J. Sattar, “Robot Communication Via Motion: A Study on Modalities for Robot-to-Human Communication in the Field,” ACM Transactions on Human-Robot Interaction (THRI), vol. 11, no. 2, p. 1–40, 2022.
- G. Dudek, J. Sattar, and A. Xu, “A Visual Language for Robot Control and Programming: A Human-Interface Study,” in Proceedings of the International Conference on Robotics and Automation ICRA, (Rome, Italy), pp. 2507–2513, April 2007.
- A. Xu, G. Dudek, and J. Sattar, “A Natural Gesture Interface for Operating Robotic Systems,” in Proceedings of the IEEE International Conference on Robotics and Automation, ICRA, (Pasadena, California), pp. 3557–3563, May 2008.
- A. Sagitov, K. Shabalina, L. Sabirova, H. Li, and E. Magid, “ARTag, AprilTag and CALTag Fiducial Marker Systems: Comparison in a Presence of Partial Marker Occlusion and Rotation.,” in ICINCO (2), p. 182–191, 2017.
- H. Buelow and A. Birk, “Gesture-recognition as basis for a human robot interface (hri) on a auv,” in OCEANS’11 MTS/IEEE KONA, pp. 1–9, Sept 2011.
- D. Chiarella, M. Bibuli, G. Bruzzone, M. Caccia, A. Ranieri, E. Zereik, L. Marconi, and P. Cutugno, “A novel gesture-based language for underwater human–robot interaction,” Journal of Marine Science and Engineering, vol. 6, no. 3, p. 91, 2018.
- M. J. Islam, M. Ho, and J. Sattar, “Understanding human motion and gestures for underwater human–robot collaboration,” Journal of Field Robotics, vol. 36, no. 5, p. 851–873, 2019.
- A. Gomez Chavez, A. Ranieri, D. Chiarella, and A. Birk, “Underwater Vision-Based Gesture Recognition: A Robustness Validation for Safe Human–Robot Interaction,” IEEE Robotics and Automation Magazine, vol. 28, no. 3, pp. 67–78, 2021.
- B. Verzijlenberg and M. Jenkin, “Swimming with Robots: Human Robot Communication at Depth,” in 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 4023–4028, 2010.
- M. Bernardi, C. Cardia, P. Gjanci, A. Monterubbiano, C. Petrioli, L. Picari, and D. Spaccini, “The Diver System: Multimedia Communication and Localization Using Underwater Acoustic Networks,” in 2019 IEEE 20th International Symposium on “A World of Wireless, Mobile and Multimedia Networks” (WoWMoM), pp. 1–8, 2019.
- MIT press, 2016.
- B. Xiao, H. Wu, and Y. Wei, “Simple baselines for human pose estimation and tracking,” in Eur. Conf. Comput. Vis., p. 466–481, 2018.
- H. Dai, H. Shi, W. Liu, L. Wang, Y. Liu, and T. Mei, “FasterPose: A Faster Simple Baseline for Human Pose Estimation,” ACM TOMM, vol. 18, no. 4, p. 1–16, 2022.
- Z. Cao, T. Simon, S.-E. Wei, and Y. Sheikh, “Realtime multi-person 2D pose estimation using part affinity fields,” in IEEE Conf. Comput. Vis. Pattern Recog., p. 7291–7299, 2017.
- Z. Yu, J. S. Yoon, I. K. Lee, P. Venkatesh, J. Park, J. Yu, and H. S. Park, “Humbi: A large multiview dataset of human body expressions,” in IEEE Conf. Comput. Vis. Pattern Recog., p. 2990–3000, 2020.
- J. S. Yoon, Z. Yu, J. Park, and H. S. Park, “HUMBI: A Large Multiview Dataset of Human Body Expressions and Benchmark Challenge,” arXiv preprint arXiv:2110.00119, 2021.
- F. Chaumette and S. Hutchinson, “Visual servo control. I. Basic approaches,” IEEE Robotics & Automation Magazine, vol. 13, no. 4, p. 82–90, 2006.
- F. Chaumette and S. Hutchinson, “Visual servoing and visual tracking,” 2008.
- F. Chaumette, S. Hutchinson, and P. Corke, “Visual servoing,” in Springer Handbook of Robotics, p. 841–866, Springer, 2016.
- A. Mathis, P. Mamidanna, K. M. Cury, T. Abe, V. N. Murthy, M. W. Mathis, and M. Bethge, “DeepLabCut: markerless pose estimation of user-defined body parts with deep learning,” Nature neuroscience, vol. 21, no. 9, p. 1281–1289, 2018.
- T. Nath, A. Mathis, A. C. Chen, A. Patel, M. Bethge, and M. W. Mathis, “Using DeepLabCut for 3D markerless pose estimation across species and behaviors,” Nature protocols, vol. 14, no. 7, p. 2152–2176, 2019.
- W. Kabsch, “A Discussion of the Solution for the Best Rotation to Relate Two Sets of Vectors,” Acta Crystallographica Section A, vol. 34, pp. 827–828, Sep 1978.