DaFoEs: Mixing Datasets towards the generalization of vision-state deep-learning Force Estimation in Minimally Invasive Robotic Surgery (2401.09239v1)
Abstract: Precisely determining the contact force during safe interaction in Minimally Invasive Robotic Surgery (MIRS) is still an open research challenge. Inspired by post-operative qualitative analysis from surgical videos, the use of cross-modality data driven deep neural network models has been one of the newest approaches to predict sensorless force trends. However, these methods required for large and variable datasets which are not currently available. In this paper, we present a new vision-haptic dataset (DaFoEs) with variable soft environments for the training of deep neural models. In order to reduce the bias from a single dataset, we present a pipeline to generalize different vision and state data inputs for mixed dataset training, using a previously validated dataset with different setup. Finally, we present a variable encoder-decoder architecture to predict the forces done by the laparoscopic tool using single input or sequence of inputs. For input sequence, we use a recurrent decoder, named with the prefix R, and a new temporal sampling to represent the acceleration of the tool. During our training, we demonstrate that single dataset training tends to overfit to the training data domain, but has difficulties on translating the results across new domains. However, dataset mixing presents a good translation with a mean relative estimated force error of 5% and 12% for the recurrent and non-recurrent models respectively. Our method, also marginally increase the effectiveness of transformers for force estimation up to a maximum of ~15%, as the volume of available data is increase by 150%. In conclusion, we demonstrate that mixing experimental set ups for vision-state force estimation in MIRS is a possible approach towards the general solution of the problem.
- L. Bahar, Y. Sharon, and I. Nisky, “Surgeon-Centered Analysis of Robot-Assisted Needle Driving Under Different Force Feedback Conditions,” Frontiers in Neurorobotics, vol. 13, p. 108, 2020.
- A. Abiri, J. Pensa, A. Tao, J. Ma, Y.-Y. Juo, S. J. Askari, J. Bisley, J. Rosen, E. P. Dutson, and W. S. Grundfest, “Multi-Modal Haptic Feedback for Grip Force Reduction in Robotic Surgery,” Scientific Reports, vol. 9, p. 5016, Mar. 2019. Number: 1 Publisher: Nature Publishing Group.
- R. V. Patel, S. F. Atashzar, and M. Tavakoli, “Haptic Feedback and Force-Based Teleoperation in Surgical Robotics,” Proceedings of the IEEE, vol. 110, pp. 1012–1027, July 2022. Conference Name: Proceedings of the IEEE.
- F. Piqué, M. N. Boushaki, M. Brancadoro, E. De Momi, and A. Menciassi, “Dynamic Modeling of the Da Vinci Research Kit Arm for the Estimation of Interaction Wrench,” in 2019 International Symposium on Medical Robotics (ISMR), pp. 1–7, Apr. 2019.
- H. Talbot, N. Haouchine, I. Peterlik, J. Dequidt, C. Duriez, H. Delingette, and S. Cotin, “Surgery Training, Planning and Guidance Using the SOFA Framework,” in Eurographics, (Zurich, Switzerland), May 2015.
- Y. Gao, S. S. Vedula, C. E. Reiley, N. Ahmidi, B. Varadarajan, H. C. Lin, L. Tao, L. Zappella, B. Béjar, D. D. Yuh, et al., “Jhu-isi gesture and skill assessment working set (jigsaws): A surgical activity dataset for human motion modeling,” in MICCAI workshop: M2cai, vol. 3, 2014.
- Z. Chua, A. M. Jarc, and A. M. Okamura, “Toward Force Estimation in Robot-Assisted Surgery using Deep Learning with Vision and Robot State,” in 2021 IEEE International Conference on Robotics and Automation (ICRA), pp. 12335–12341, May 2021. ISSN: 2577-087X.
- A. Marban, V. Srinivasan, W. Samek, J. Fernández, and A. Casals, “A recurrent convolutional neural network approach for sensorless force estimation in robotic surgery,” Biomedical Signal Processing and Control, vol. 50, pp. 134–150, Apr. 2019.
- W. Liu, A. Pickett, K. Huang, and Y.-H. Su, “Camera Configuration Models for Machine Vision Based Force Estimation in Robot-Assisted Soft Body Manipulation,” in 2022 International Symposium on Medical Robotics (ISMR), pp. 1–8, Apr. 2022. ISSN: 2771-9049.
- P. V. Sabique, P. Ganesh, and R. Sivaramakrishnan, “Stereovision based force estimation with stiffness mapping in surgical tool insertion using recurrent neural network,” The Journal of Supercomputing, vol. 78, pp. 14648–14679, Aug. 2022.
- P. V. Sabique, G. Pasupathy, and S. Ramachandran, “A data driven recurrent neural network approach for reproduction of variable visuo-haptic force feedback in surgical tool insertion,” Expert Systems with Applications, vol. 238, p. 122221, Mar. 2024.
- P. V. Sabique, G. Pasupathy, S. Ramachandran, and G. Shanmugasundar, “Investigating the influence of dimensionality reduction on force estimation in robotic-assisted surgery using recurrent and convolutional networks,” Engineering Applications of Artificial Intelligence, vol. 126, p. 107045, Nov. 2023.
- A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly, J. Uszkoreit, and N. Houlsby, “An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale,” June 2021. arXiv:2010.11929 [cs].
- M. Fayyaz, S. A. Koohpayegani, F. R. Jafari, S. Sengupta, H. R. V. Joze, E. Sommerlade, H. Pirsiavash, and J. Gall, “Adaptive Token Sampling For Efficient Vision Transformers,” July 2022. arXiv:2111.15667 [cs].
- A. Torralba and A. A. Efros, “Unbiased look at dataset bias,” in CVPR 2011, pp. 1521–1528, June 2011. ISSN: 1063-6919.
- R. Ranftl, K. Lasinger, D. Hafner, K. Schindler, and V. Koltun, “Towards Robust Monocular Depth Estimation: Mixing Datasets for Zero-shot Cross-dataset Transfer,” Aug. 2020. arXiv:1907.01341 [cs].
- P. Kazanzides, Z. Chen, A. Deguet, G. S. Fischer, R. H. Taylor, and S. P. DiMaio, “An open-source research kit for the da Vinci® Surgical System,” in 2014 IEEE International Conference on Robotics and Automation (ICRA), pp. 6434–6439, May 2014. ISSN: 1050-4729.
- S. Longpre, Y. Wang, and C. DuBois, “How Effective is Task-Agnostic Data Augmentation for Pretrained Transformers?,” Oct. 2020. arXiv:2010.01764 [cs, stat].
- D.-H. Lee, K.-S. Kwak, and S.-C. Lim, “A Neural Network-based Suture-tension Estimation Method Using Spatio-temporal Features of Visual Information and Robot-state Information for Robot-assisted Surgery,” International Journal of Control, Automation and Systems, vol. 21, pp. 4032–4040, Nov. 2023.
- Y.-E. Lee, H. M. Husin, M.-P. Forte, S.-W. Lee, and K. J. Kuchenbecker, “Learning to Estimate Palpation Forces in Robotic Surgery From Visual-Inertial Data,” IEEE Transactions on Medical Robotics and Bionics, vol. 5, pp. 496–506, Aug. 2023.