A Self-supervised Contrastive Learning Method for Grasp Outcomes Prediction
Abstract: In this paper, we investigate the effectiveness of contrastive learning methods for predicting grasp outcomes in an unsupervised manner. By utilizing a publicly available dataset, we demonstrate that contrastive learning methods perform well on the task of grasp outcomes prediction. Specifically, the dynamic-dictionary-based method with the momentum updating technique achieves a satisfactory accuracy of 81.83% using data from one single tactile sensor, outperforming other unsupervised methods. Our results reveal the potential of contrastive learning methods for applications in the field of robot grasping and highlight the importance of accurate grasp prediction for achieving stable grasps.
- J. Kwiatkowski, D. Cockburn, and V. Duchaine, “Grasp stability assessment through the fusion of proprioception and tactile signals using convolutional neural networks,” in 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, 2017, pp. 286–292.
- S. Fang, Z. Yi, T. Mi, Z. Zhou, C. Ye, W. Shang, T. Xu, and X. Wu, “Tactonet: Tactile ordinal network based on unimodal probability for object hardness classification,” IEEE Transactions on Automation Science and Engineering, 2022.
- D. Guo, F. Sun, B. Fang, C. Yang, and N. Xi, “Robotic grasping using visual and tactile sensing,” Information Sciences, vol. 417, pp. 274–286, 2017.
- M. Sundermeyer, A. Mousavian, R. Triebel, and D. Fox, “Contact-graspnet: Efficient 6-dof grasp generation in cluttered scenes,” in 2021 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 2021, pp. 13 438–13 444.
- J. Mahler, M. Matl, V. Satish, M. Danielczuk, B. DeRose, S. McKinley, and K. Goldberg, “Learning ambidextrous robot grasping policies,” Science Robotics, vol. 4, no. 26, p. eaau4984, 2019.
- Z. Yi, T. Xu, W. Shang, and X. Wu, “Touch modality identification with tensorial tactile signals: A kernel-based approach,” IEEE Transactions on Automation Science and Engineering, vol. 19, no. 2, pp. 959–968, 2021.
- B. R. Phelan, K. I. Ranney, K. A. Gallagher, J. T. Clark, K. D. Sherbondy, and R. M. Narayanan, “Design of ultrawideband stepped-frequency radar for imaging of obscured targets,” IEEE Sensors Journal, vol. 17, no. 14, pp. 4435–4446, 2017.
- C. Wang, X. Zang, H. Zhang, H. Chen, Z. Lin, and J. Zhao, “Status identification and object in-hand reorientation using force/torque sensors,” IEEE Sensors Journal, vol. 21, no. 18, pp. 20 694–20 703, 2021.
- A. Tulbure and B. Bäuml, “Superhuman performance in tactile material classification and differentiation with a flexible pressure-sensitive skin,” in 2018 IEEE-RAS 18th International Conference on Humanoid Robots (Humanoids). IEEE, 2018, pp. 1–9.
- S. S. Baishya and B. Bäuml, “Robust material classification with a tactile skin using deep learning,” in 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, 2016, pp. 8–15.
- Z. Zhou, B. Huang, R. Zhang, M. Yin, C. Liu, Y. Liu, Z. Yi, and X. Wu, “Methods to recognize depth of hard inclusions in soft tissue using ordinal classification for robotic palpation,” IEEE Transactions on Instrumentation and Measurement, vol. 71, pp. 1–12, 2022.
- S. Huang and H. Wu, “Texture recognition based on perception data from a bionic tactile sensor,” Sensors, vol. 21, no. 15, p. 5224, 2021.
- T. Mi, D. Que, S. Fang, Z. Zhou, C. Ye, C. Liu, Z. Yi, and X. Wu, “Tactile grasp stability classification based on graph convolutional networks,” in 2021 IEEE International Conference on Real-time Computing and Robotics (RCAR). IEEE, 2021, pp. 875–880.
- D. Cockbum, J.-P. Roberge, A. Maslyczyk, V. Duchaine et al., “Grasp stability assessment through unsupervised feature learning of tactile images,” in 2017 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 2017, pp. 2238–2244.
- R. Calandra, A. Owens, M. Upadhyaya, W. Yuan, J. Lin, E. H. Adelson, and S. Levine, “The feeling of success: Does touch sensing help predict grasp outcomes?” in Conference on Robot Learning. PMLR, 2017, pp. 314–323.
- S. Cui, R. Wang, J. Wei, J. Hu, and S. Wang, “Self-attention based visual-tactile fusion learning for predicting grasp outcomes,” IEEE Robotics and Automation Letters, vol. 5, no. 4, pp. 5827–5834, 2020.
- C. Yang, P. Du, F. Sun, B. Fang, and J. Zhou, “Predict robot grasp outcomes based on multi-modal information,” in 2018 IEEE International Conference on Robotics and Biomimetics (ROBIO). IEEE, 2018, pp. 1563–1568.
- S. Gidaris, P. Singh, and N. Komodakis, “Unsupervised representation learning by predicting image rotations,” arXiv preprint arXiv:1803.07728, 2018.
- A. v. d. Oord, Y. Li, and O. Vinyals, “Representation learning with contrastive predictive coding,” arXiv preprint arXiv:1807.03748, 2018.
- Z. Wu, Y. Xiong, S. X. Yu, and D. Lin, “Unsupervised feature learning via non-parametric instance discrimination,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 3733–3742.
- K. He, H. Fan, Y. Wu, S. Xie, and R. Girshick, “Momentum contrast for unsupervised visual representation learning,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 9729–9738.
- W. Yuan, S. Dong, and E. H. Adelson, “Gelsight: High-resolution robot tactile sensors for estimating geometry and force,” Sensors, vol. 17, no. 12, p. 2762, 2017.
- X. Chen, H. Fan, R. Girshick, and K. He, “Improved baselines with momentum contrastive learning,” arXiv preprint arXiv:2003.04297, 2020.
- T. Chen, S. Kornblith, M. Norouzi, and G. Hinton, “A simple framework for contrastive learning of visual representations,” in International conference on machine learning. PMLR, 2020, pp. 1597–1607.
- Z. Wu, Y. Xiong, S. X. Yu, and D. Lin, “Unsupervised feature learning via non-parametric instance discrimination,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2018.
- M. Ye, X. Zhang, P. C. Yuen, and S.-F. Chang, “Unsupervised embedding learning via invariant and spreading instance feature,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2019.
- K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778.
- K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” arXiv preprint arXiv:1409.1556, 2014.
- E. Hoffer and N. Ailon, “Deep metric learning using triplet network,” in International workshop on similarity-based pattern recognition. Springer, 2015, pp. 84–92.
- G. E. Hinton and R. R. Salakhutdinov, “Reducing the dimensionality of data with neural networks,” science, vol. 313, no. 5786, pp. 504–507, 2006.
- I. Loshchilov and F. Hutter, “Sgdr: Stochastic gradient descent with warm restarts,” arXiv preprint arXiv:1608.03983, 2016.
- F. Schroff, D. Kalenichenko, and J. Philbin, “Facenet: A unified embedding for face recognition and clustering,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 815–823.
- U. Lall and A. Sharma, “A nearest neighbor bootstrap for resampling hydrologic time series,” Water resources research, vol. 32, no. 3, pp. 679–693, 1996.
Sponsor
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.