Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
125 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Self-Supervised Learning of Visual Servoing for Low-Rigidity Robots Considering Temporal Body Changes (2405.11798v1)

Published 20 May 2024 in cs.RO

Abstract: In this study, we investigate object grasping by visual servoing in a low-rigidity robot. It is difficult for a low-rigidity robot to handle its own body as intended compared to a rigid robot, and calibration between vision and body takes some time. In addition, the robot must constantly adapt to changes in its body, such as the change in camera position and change in joints due to aging. Therefore, we develop a method for a low-rigidity robot to autonomously learn visual servoing of its body. We also develop a mechanism that can adaptively change its visual servoing according to temporal body changes. We apply our method to a low-rigidity 6-axis arm, MyCobot, and confirm its effectiveness by conducting object grasping experiments based on visual servoing.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (22)
  1. Y. Shirai and H. Inoue, “Guiding a robot by visual feedback in assembling tasks,” Pattern Recognition, vol. 5, no. 2, pp. 99–108, 1973.
  2. B. Espiau, F. Chaumette, and P. Rives, “A new approach to visual servoing in robotics,” IEEE Transactions on Robotics and Automation, vol. 8, no. 3, pp. 313–326, 1992.
  3. T. Lampe and M. Riedmiller, “Acquiring visual servoing reaching and grasping skills using neural reinforcement learning,” in Proceedings of the 2013 International Joint Conference on Neural Networks, 2013, pp. 1–8.
  4. A. X. Lee, S. Levine, and P. Abbeel, “Learning Visual Servoing with Deep Features and Fitted Q-Iteration,” in Proceedings of the 5th International Conference on Learning Representations, 2017, pp. 1–20.
  5. T. Zhang, Z. McCarthy, O. Jow, D. Lee, X. Chen, K. Goldberg, and P. Abbeel, “Deep Imitation Learning for Complex Manipulation Tasks from Virtual Reality Teleoperation,” in Proceedings of the 2018 IEEE International Conference on Robotics and Automation, 2018, pp. 5628–5635.
  6. N. Saito, T. Ogata, S. Funabashi, H. Mori, and S. Sugano, “How to Select and Use Tools? : Active Perception of Target Objects Using Multimodal Deep Learning,” IEEE Robotics and Automation Letters, vol. 6, no. 2, pp. 2517–2524, 2021.
  7. H. Wang, B. Yang, Y. Liu, W. Chen, X. Liang, and R. Pfeifer, “Visual Servoing of Soft Robot Manipulator in Constrained Environments With an Adaptive Controller,” IEEE/ASME Transactions on Mechatronics, vol. 22, no. 1, pp. 41–50, 2017.
  8. P. Chatelain, A. Krupa, and N. Navab, “3D ultrasound-guided robotic steering of a flexible needle via visual servoing,” in Proceedings of the 2015 IEEE International Conference on Robotics and Automation, 2015, pp. 2250–2255.
  9. K. Kawaharazuka, K. Tsuzuki, M. Onitsuka, Y. Asano, K. Okada, K. Kawasaki, and M. Inaba, “Object Recognition, Dynamic Contact Simulation, Detection, and Control of the Flexible Musculoskeletal Hand Using a Recurrent Neural Network With Parametric Bias,” IEEE Robotics and Automation Letters, vol. 5, no. 3, pp. 4580–4587, 2020.
  10. C. Choi, W. Schwarting, J. DelPreto, and D. Rus, “Learning Object Grasping for Soft Robot Hands,” IEEE Robotics and Automation Letters, vol. 3, no. 3, pp. 2370–2377, 2018.
  11. M. Kanamura, K. Suzuki, Y. Suga, and T. Ogata, “Development of a Basic Educational Kit for Robotic System with Deep Neural Networks,” Sensors, vol. 21, no. 11, pp. 1–21, 2021.
  12. J. Tani, “Self-organization of behavioral primitives as multiple attractor dynamics: a robot experiment,” in Proceedings of the 2002 International Joint Conference on Neural Networks, 2002, pp. 489–494.
  13. T. Ogata, H. Ohba, J. Tani, K. Komatani, and H. G. Okuno, “Extracting multi-modal dynamics of objects using RNNPB,” in Proceedings of the 2005 IEEE/RSJ International Conference on Intelligent Robots and Systems, 2005, pp. 966–971.
  14. R. Yokoya, T. Ogata, J. Tani, K. Komatani, and H. G. Okuno, “Experience Based Imitation Using RNNPB,” in Proceedings of the 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems, 2006, pp. 3669–3674.
  15. G. E. Hinton and R. R. Salakhutdinov, “Reducing the Dimensionality of Data with Neural Networks,” Science, vol. 313, no. 5786, pp. 504–507, 2006.
  16. K. Kawaharazuka, S. Makino, K. Tsuzuki, M. Onitsuka, Y. Nagamatsu, K. Shinjo, T. Makabe, Y. Asano, K. Okada, K. Kawasaki, and M. Inaba, “Component Modularized Design of Musculoskeletal Humanoid Platform Musashi to Investigate Learning Control Systems,” in Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems, 2019, pp. 7294–7301.
  17. S. Hochreiter and J. Schmidhuber, “Long short-term memory,” Neural computation, vol. 9, no. 8, pp. 1735–1780, 1997.
  18. D. P. Kingma and J. Ba, “Adam: A Method for Stochastic Optimization,” in Proceedings of the 3rd International Conference on Learning Representations, 2015, pp. 1–15.
  19. S. Ioffe and C. Szegedy, “Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift,” in Proceedings of the 32nd International Conference on Machine Learning, 2015, pp. 448–456.
  20. V. Nair and G. E. Hinton, “Rectified Linear Units Improve Restricted Boltzmann Machines,” in Proceedings of the 27th International Conference on Machine Learning, 2010, pp. 807–814.
  21. C. Laschi, M. Cianchetti, B. Mazzolai, L. Margheri, M. Follador, and P. Dario, “Soft Robot Arm Inspired by the Octopus,” Advanced Robotics, vol. 26, no. 7, pp. 709–727, 2012.
  22. P. Florence, C. Lynch, A. Zeng, O. Ramirez, A. Wahid, L. Downs, A. Wong, J. Lee, I. Mordatch, and J. Tompson, “Implicit Behavioral Cloning,” in Proceedings of the 2021 Conference on Robot Learning, 2021.
Citations (4)

Summary

  • The paper presents a novel self-supervised approach that integrates parametric bias into a visual servoing network to adapt to temporal changes.
  • The method uses repetitive, self-directed object manipulation tasks to enhance calibration and efficiency in low-rigidity robots.
  • Experimental results confirm that the VSNPB model maintains high grasp accuracy despite mechanical variations, underlining its potential for flexible robotics.

Self-Supervised Learning of Visual Servoing for Low-Rigidity Robots

The paper delineated in "Self-Supervised Learning of Visual Servoing for Low-Rigidity Robots Considering Temporal Body Changes" is predicated upon developing methodologies to enhance robotic flexibility, specifically for systems composed of low-rigidity materials. The system facilitates autonomous learning for low-rigidity robots, which are commonly plagued by calibration challenges owing to body compliance. This compliance makes the conventional model training and application of robot perception to action a complex affair, requiring constant recalibration to accommodate shifts in the robot's kinematic chain due to mechanical wear and tear or even operational shifts.

The authors' emphasis is on leveraging the inherent ability of low-rigidity robots to model their body's visual servoing mechanisms autonomously. They present a paradigm whereby a robot, MyCobot in this case, can engage in self-supervised learning, thus negating the need for human intervention in calibration and teaching of the robot's tasks, which is a merit considering operational scalability and efficiency. The paper introduces parametric bias (PB) to the learning mechanism, allowing the model to adapt dynamically to temporal changes in the robot's body, promoting the ability to respond effectively to cumulative structural alterations.

A pivotal facet of the paper is the novel data collection strategy wherein the robot executes a self-directed task of placing and picking up an object without external demonstrations or reliance on reinforcement learning. This strategy of using repetitive task execution as a learning mechanism allows the robot to consistently and dynamically align its perceived visual data with actual physical transformations. The robot can perform accurate object grasping, leveraging the high repeatability of its movements to appreciate the temporal changes in its joint configurations and visual perceptions.

In detail, the research employs a visual servoing network with parametric bias (VSNPB), an iterative closed-loop system integrating the robot's joint states and visual inputs to refine function over time. This system is handily adaptable to low-rigidity configurations due to its capacity to embed dynamic modeling errors arising from temporal body changes directly into the bias input, facilitating real-time adaptation.

The experimental results underscore the capacity of the VSNPB model to manage evolving body dynamics efficiently. With multiple body states (comprising alignment variations and camera position shifts), the network exhibits resilience in maintaining grasp accuracy amidst perturbations in the physical and sensory environment. For each test object, the success rates depicted a strong correlation with the usage of the correct versus incorrect parametric biases, validating the model's premises about the importance of synchronous sensor-actuator state recognition.

Potential applications for this research extend into realms where low-cost, lightweight robotic implementations are prioritized or indispensable, such as in service robotics, distributed sensors in flexible environments, and adaptive manufacturing processes. The extrapolation to more complex, iterative tasks or multi-object manipulation underscores a potent avenue for future research.

This research addresses robotic vision and control's agile adaption, posing a significant contribution towards autonomous calibration processes within robotic systems, specifically those limited by structural pliability and where multi-sensor integration and real-time adaptability bear significant applications. Future work may focus on the exploratory application of this method to a wider range of robots with variable structural configurations and environmental influences to further test and refine the efficacy and robustness of the parametric bias framework introduced herein.

X Twitter Logo Streamline Icon: https://streamlinehq.com
Youtube Logo Streamline Icon: https://streamlinehq.com