Coaching a Robotic Sonographer: Learning Robotic Ultrasound with Sparse Expert's Feedback (2409.02337v1)
Abstract: Ultrasound is widely employed for clinical intervention and diagnosis, due to its advantages of offering non-invasive, radiation-free, and real-time imaging. However, the accessibility of this dexterous procedure is limited due to the substantial training and expertise required of operators. The robotic ultrasound (RUS) offers a viable solution to address this limitation; nonetheless, achieving human-level proficiency remains challenging. Learning from demonstrations (LfD) methods have been explored in RUS, which learns the policy prior from a dataset of offline demonstrations to encode the mental model of the expert sonographer. However, active engagement of experts, i.e. Coaching, during the training of RUS has not been explored thus far. Coaching is known for enhancing efficiency and performance in human training. This paper proposes a coaching framework for RUS to amplify its performance. The framework combines DRL (self-supervised practice) with sparse expert's feedback through coaching. The DRL employs an off-policy Soft Actor-Critic (SAC) network, with a reward based on image quality rating. The coaching by experts is modeled as a Partially Observable Markov Decision Process (POMDP), which updates the policy parameters based on the correction by the expert. The validation study on phantoms showed that coaching increases the learning rate by $25\%$ and the number of high-quality image acquisition by $74.5\%$.
- J. C. Carr, G. R. Gerstner, C. C. Voskuil, J. E. Harden, D. Dunnick, K. M. Badillo, J. I. Pagan, K. K. Harmon, R. M. Girts, J. P. Beausejour et al., “The influence of sonographer experience on skeletal muscle image acquisition and analysis,” Journal of functional morphology and kinesiology, vol. 6, no. 4, p. 91, 2021.
- D. Raina, H. Singh, S. K. Saha, C. Arora, A. Agarwal, S. Chandrashekhara, K. Rangarajan, and S. Nandi, “Comprehensive telerobotic ultrasound system for abdominal imaging: Development and in-vivo feasibility study,” in 2021 International Symposium on Medical Robotics (ISMR). IEEE, 2021, pp. 1–7.
- S. H. Chandrashekhara, K. Rangarajan, A. Agrawal, S. Thulkar, S. Gamanagatti, D. Raina, S. K. Saha, and C. Arora, “Robotic ultrasound: An initial feasibility study,” World Journal of Methodology, vol. 12, no. 4, p. 274, 2022.
- Z. Jiang, S. E. Salcudean, and N. Navab, “Robotic ultrasound imaging: State-of-the-art and future perspectives,” Medical image analysis, p. 102878, 2023.
- M. V. Balakuntala, V. L. Venkatesh, J. P. Bindu, R. M. Voyles, and J. Wachs, “Extending policy from one-shot learning through coaching,” in 2019 28th IEEE International Conference on Robot and Human Interactive Communication (RO-MAN). IEEE, 2019, pp. 1–7.
- M. B. I. Pamies, M. T. Villasevil, Z. Wang, S. Desai, P. Agrawal, and A. Gupta, “Autonomous robotic reinforcement learning with asynchronous human feedback,” in 7th Annual Conference on Robot Learning, 2023.
- K. Li, J. Wang, Y. Xu, H. Qin, D. Liu, L. Liu, and M. Q.-H. Meng, “Autonomous navigation of an ultrasound probe towards standard scan planes with deep reinforcement learning,” in 2021 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 2021, pp. 8302–8308.
- G. Ning, X. Zhang, and H. Liao, “Autonomic robotic ultrasound imaging system based on reinforcement learning,” IEEE Transactions on Biomedical Engineering, vol. 68, no. 9, pp. 2787–2797, 2021.
- D. Raina, D. Ntentia, S. Chandrashekhara, R. Voyles, and S. K. Saha, “Expert-agnostic ultrasound image quality assessment using deep variational clustering,” in 2023 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 2023, pp. 2717–2723.
- Z. Jiang, Y. Bi, M. Zhou, Y. Hu, M. Burke, and N. Navab, “Intelligent robotic sonographer: Mutual information-based disentangled reward learning from few demonstrations,” The International Journal of Robotics Research, p. 02783649231223547, 2023.
- R. Mebarki, A. Krupa, and F. Chaumette, “2-d ultrasound probe complete guidance by visual servoing using image moments,” IEEE Transactions on Robotics, vol. 26, no. 2, pp. 296–306, 2010.
- C. Nadeau, A. Krupa, J. Petr, and C. Barillot, “Moments-based ultrasound visual servoing: From a mono-to multiplane approach,” IEEE Transactions on Robotics, vol. 32, no. 6, pp. 1558–1564, 2016.
- C. Hennersperger, B. Fuerst, S. Virga, O. Zettinig, B. Frisch, T. Neff, and N. Navab, “Towards mri-based autonomous robotic us acquisitions: a first feasibility study,” IEEE transactions on medical imaging, vol. 36, no. 2, pp. 538–548, 2016.
- L. Al-Zogbi, V. Singh, B. Teixeira, A. Ahuja, P. S. Bagherzadeh, A. Kapoor, H. Saeidi, T. Fleiter, and A. Krieger, “Autonomous robotic point-of-care ultrasound imaging for monitoring of covid-19–induced pulmonary diseases,” Frontiers in Robotics and AI, vol. 8, p. 645756, 2021.
- G. P. Mylonas, P. Giataganas, M. Chaudery, V. Vitiello, A. Darzi, and G.-Z. Yang, “Autonomous efast ultrasound scanning by a robotic manipulator using learning from demonstrations,” in 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems. IEEE, 2013, pp. 3251–3256.
- M. Li and X. Deng, “Learning robotic ultrasound skills from human demonstrations,” in Cognitive Robotics. IntechOpen, 2022.
- D. Raina, S. Chandrashekhara, R. Voyles, J. Wachs, and S. K. Saha, “Robotic sonographer: Autonomous robotic ultrasound using domain expertise in bayesian optimization,” in 2023 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 2023, pp. 6909–6915.
- ——, “Deep kernel and image quality estimators for optimizing robotic ultrasound controller using bayesian optimization,” in 2023 International Symposium on Medical Robotics (ISMR). IEEE, 2023, pp. 1–7.
- D. Raina, A. Mathur, R. M. Voyles, J. Wachs, S. Chandrashekhara, and S. K. Saha, “Rusopt: Robotic ultrasound probe normalization with bayesian optimization for in-plane and out-plane scanning,” in 2023 IEEE 19th International Conference on Automation Science and Engineering (CASE). IEEE, 2023, pp. 1–7.
- M. Burke, K. Lu, D. Angelov, A. Straižys, C. Innes, K. Subr, and S. Ramamoorthy, “Learning rewards from exploratory demonstrations using probabilistic temporal ranking,” Autonomous Robots, vol. 47, no. 6, pp. 733–751, 2023.
- A. L. Thomaz, G. Hoffman, and C. Breazeal, “Real-time interactive reinforcement learning for robots,” in AAAI 2005 workshop on human comprehensible machine learning, vol. 3, no. 3.7, 2005, p. 1.
- J. MacGlashan, M. K. Ho, R. Loftin, B. Peng, G. Wang, D. L. Roberts, M. E. Taylor, and M. L. Littman, “Interactive learning from policy-dependent human feedback,” in International conference on machine learning. PMLR, 2017, pp. 2285–2294.
- E. Biyik and D. Sadigh, “Batch active preference-based learning of reward functions,” in Conference on robot learning. PMLR, 2018, pp. 519–528.
- S. Griffith, K. Subramanian, J. Scholz, C. L. Isbell, and A. L. Thomaz, “Policy shaping: Integrating human feedback with reinforcement learning,” Advances in neural information processing systems, vol. 26, 2013.
- D. P. Losey and M. K. O’Malley, “Learning the correct robot trajectory in real-time from physical human interactions,” ACM Transactions on Human-Robot Interaction (THRI), vol. 9, no. 1, pp. 1–19, 2019.
- A. Bobu, A. Bajcsy, J. F. Fisac, S. Deglurkar, and A. D. Dragan, “Quantifying hypothesis space misspecification in learning from human–robot demonstrations and physical corrections,” IEEE Transactions on Robotics, vol. 36, no. 3, pp. 835–854, 2020.
- M. Li, A. Canberk, D. P. Losey, and D. Sadigh, “Learning human objectives from sequences of physical corrections,” in 2021 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 2021, pp. 2877–2883.
- S. Javdani, H. Admoni, S. Pellegrinelli, S. S. Srinivasa, and J. A. Bagnell, “Shared autonomy via hindsight optimization for teleoperation and teaming,” The International Journal of Robotics Research, vol. 37, no. 7, pp. 717–742, 2018.
- T. Haarnoja, A. Zhou, P. Abbeel, and S. Levine, “Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor,” in International conference on machine learning. PMLR, 2018, pp. 1861–1870.
- A. Singh, L. Yang, C. Finn, and S. Levine, “End-to-end robotic reinforcement learning without reward engineering,” Robotics: Science and Systems XV, 2019.
- D. Raina, S. Chandrashekhara, R. Voyles, J. Wachs, and S. K. Saha, “Deep learning model for quality assessment of urinary bladder ultrasound images using multi-scale and higher-order processing,” IEEE Transactions on Ultrasonics, Ferroelectrics, and Frequency Control, 2024.
- M. Chen, E. Frazzoli, D. Hsu, and W. S. Lee, “Pomdp-lite for robust robot planning under uncertainty,” in 2016 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 2016, pp. 5427–5433.
- D. P. Losey, A. Bajcsy, M. K. O’Malley, and A. D. Dragan, “Physical interaction as communication: Learning robot objectives online from human corrections,” The International Journal of Robotics Research, vol. 41, no. 1, pp. 20–44, 2022.
Sponsor
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.