Extrinsic Manipulation on a Support Plane by Learning Regrasping (2210.05349v3)
Abstract: Extrinsic manipulation, a technique that enables robots to leverage extrinsic resources for object manipulation, presents practical yet challenging scenarios. Particularly in the context of extrinsic manipulation on a supporting plane, regrasping becomes essential for achieving the desired final object poses. This process involves sequential operation steps and stable placements of objects, which provide grasp space for the robot. To address this challenge, we focus on predicting diverse placements of objects on the plane using deep neural networks. A framework that comprises orientation generation, placement refinement, and placement discrimination stages is proposed, leveraging point clouds to obtain precise and diverse stable placements. To facilitate training, a large-scale dataset is constructed, encompassing stable object placements and contact information between objects. Through extensive experiments, our approach is demonstrated to outperform the start-of-the-art, achieving an accuracy rate of 90.4\% and a diversity rate of 81.3\% in predicted placements. Furthermore, we validate the effectiveness of our approach through real-robot experiments, demonstrating its capability to compute sequential pick-and-place steps based on the predicted placements for regrasping objects to goal poses that are not readily attainable within a single step. Videos and dataset are available at https://sites.google.com/view/pmvlr2022/.
- N. C. Dafle, A. Rodriguez, R. Paolini, B. Tang, S. S. Srinivasa, M. Erdmann, M. T. Mason, I. Lundberg, H. Staab, and T. Fuhlbrigge, “Extrinsic dexterity: In-hand manipulation with external forces,” in 2014 IEEE International Conference on Robotics and Automation (ICRA), pp. 1578–1585, 2014.
- W. Wan, K. Harada, and F. Kanehiro, “Preparatory manipulation planning using automatically determined single and dual arm,” IEEE Transactions on Industrial Informatics, vol. 16, no. 1, pp. 442–453, 2019.
- W. Wan and K. Harada, “Achieving high success rate in dual-arm handover using large number of candidate grasps, handover heuristics, and hierarchical search,” Advanced Robotics, vol. 30, no. 17-18, pp. 1111–1125, 2016.
- F. Suárez-Ruiz, X. Zhou, and Q.-C. Pham, “Can robots assemble an ikea chair?,” Science Robotics, vol. 3, no. 17, p. eaat6385, 2018.
- Y. Shi, C. Yuan, A. Tsitos, L. Cong, H. Hadjar, Z. Chen, and J. Zhang, “A sim-to-real learning based framework for contact-rich assembly by utilizing cyclegan and force control,” IEEE Transactions on Cognitive and Developmental Systems, 2023.
- R. Jiang, B. He, Z. Wang, Y. Zhou, S. Xu, and X. Li, “A novel simulation-reality closed-loop learning framework for autonomous robot skill learning,” IEEE Transactions on Cognitive and Developmental Systems, vol. 14, no. 4, pp. 1520–1531, 2021.
- H. Duan, P. Wang, Y. Li, D. Li, and W. Wei, “Learning human-to-robot dexterous handovers for anthropomorphic hand,” IEEE Transactions on Cognitive and Developmental Systems, 2022.
- S. Cheng, K. Mo, and L. Shao, “Learning to regrasp by learning to place,” arXiv preprint arXiv:2109.08817, 2021.
- P. Xu, H. Cheng, J. Wang, and M. Q.-H. Meng, “Learning to reorient objects with stable placements afforded by extrinsic supports,” arXiv preprint arXiv:2205.06970, 2022.
- A. Simeonov, Y. Du, B. Kim, F. R. Hogan, J. Tenenbaum, P. Agrawal, and A. Rodriguez, “A long horizon planning framework for manipulating rigid pointcloud objects,” in Conference on Robot Learning (CoRL), 2020.
- W. Wan, H. Igawa, K. Harada, H. Onda, K. Nagata, and N. Yamanobe, “A regrasp planning component for object reorientation,” Autonomous Robots, vol. 43, no. 5, pp. 1101–1115, 2019.
- J. Ma, W. Wan, K. Harada, Q. Zhu, and H. Liu, “Regrasp planning using stable object poses supported by complex structures,” IEEE Transactions on Cognitive and Developmental Systems, vol. 11, no. 2, pp. 257–269, 2018.
- K. Wada, S. James, and A. J. Davison, “Reorientbot: Learning object reorientation for specific-posed placement,” arXiv preprint arXiv:2202.11092, 2022.
- C. Paxton, C. Xie, T. Hermans, and D. Fox, “Predicting stable configurations for semantic placement of novel objects,” in Conference on Robot Learning, pp. 806–815, PMLR, 2022.
- Y. Hou, Z. Jia, and M. T. Mason, “Reorienting objects in 3d space using pivoting,” arXiv preprint arXiv:1912.02752, 2019.
- Y. You, L. Shao, T. Migimatsu, and J. Bohg, “Omnihang: Learning to hang arbitrary objects using contact point correspondences and neural collision estimation,” in 2021 IEEE International Conference on Robotics and Automation (ICRA), pp. 5921–5927, IEEE, 2021.
- X. Pang, F. Li, N. Ding, and X. Zhong, “Upright-net: Learning upright orientation for 3d point cloud,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 14911–14919, 2022.
- Y. Shi, J. Huang, X. Xu, Y. Zhang, and K. Xu, “Stablepose: Learning 6d object poses from geometrically stable patches,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 15222–15231, 2021.
- R. Newbury, K. He, A. Cosgun, and T. Drummond, “Learning to place objects onto flat surfaces in upright orientations,” IEEE Robotics and Automation Letters, vol. 6, no. 3, pp. 4377–4384, 2021.
- C. R. Qi, L. Yi, H. Su, and L. J. Guibas, “Pointnet++: Deep hierarchical feature learning on point sets in a metric space,” Advances in neural information processing systems, vol. 30, 2017.
- Z. Jiang, Y. Zhu, M. Svetlik, K. Fang, and Y. Zhu, “Synergies between affordance and geometry: 6-dof grasp detection via implicit representations,” arXiv preprint arXiv:2104.01542, 2021.
- M. Breyer, J. J. Chung, L. Ott, R. Siegwart, and J. Nieto, “Volumetric grasping network: Real-time 6 dof grasp detection in clutter,” arXiv preprint arXiv:2101.01132, 2021.
- I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial networks,” Communications of the ACM, vol. 63, no. 11, pp. 139–144, 2020.
- H. Fan, H. Su, and L. J. Guibas, “A point set generation network for 3d object reconstruction from a single image,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 605–613, 2017.
- Y. Zhou, C. Barnes, J. Lu, J. Yang, and H. Li, “On the continuity of rotation representations in neural networks,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 5745–5753, 2019.
- S. Sarabandi and F. Thomas, “A survey on the computation of quaternions from rotation matrices,” Journal of mechanisms and robotics, vol. 11, no. 2, 2019.
- J. Á. Cid and F. A. F. Tojo, “A lipschitz condition along a transversal foliation implies local uniqueness for odes,” arXiv preprint arXiv:1801.01724, 2018.
- O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in International Conference on Medical image computing and computer-assisted intervention, pp. 234–241, Springer, 2015.
- K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770–778, 2016.
- S. Kim, H.-g. Chi, X. Hu, Q. Huang, and K. Ramani, “A large-scale annotated mechanical components benchmark for classification and retrieval tasks with deep neural networks,” in Proceedings of 16th European Conference on Computer Vision (ECCV), 2020.
- E. Coumans and Y. Bai, “Pybullet, a python module for physics simulation for games, robotics and machine learning,” 2016.
- K. Fukunaga and L. Hostetler, “The estimation of the gradient of a density function, with applications in pattern recognition,” IEEE Transactions on Information Theory, vol. 21, no. 1, pp. 32–40, 1975.
- A. Paszke, S. Gross, F. Massa, A. Lerer, J. Bradbury, G. Chanan, T. Killeen, Z. Lin, N. Gimelshein, L. Antiga, et al., “Pytorch: An imperative style, high-performance deep learning library,” Advances in neural information processing systems, vol. 32, 2019.