Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Combining Self-Supervised Learning and Imitation for Vision-Based Rope Manipulation (1703.02018v1)

Published 6 Mar 2017 in cs.CV, cs.LG, and cs.RO

Abstract: Manipulation of deformable objects, such as ropes and cloth, is an important but challenging problem in robotics. We present a learning-based system where a robot takes as input a sequence of images of a human manipulating a rope from an initial to goal configuration, and outputs a sequence of actions that can reproduce the human demonstration, using only monocular images as input. To perform this task, the robot learns a pixel-level inverse dynamics model of rope manipulation directly from images in a self-supervised manner, using about 60K interactions with the rope collected autonomously by the robot. The human demonstration provides a high-level plan of what to do and the low-level inverse model is used to execute the plan. We show that by combining the high and low-level plans, the robot can successfully manipulate a rope into a variety of target shapes using only a sequence of human-provided images for direction.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Ashvin Nair (20 papers)
  2. Dian Chen (30 papers)
  3. Pulkit Agrawal (103 papers)
  4. Phillip Isola (84 papers)
  5. Pieter Abbeel (372 papers)
  6. Jitendra Malik (211 papers)
  7. Sergey Levine (531 papers)
Citations (297)

Summary

We haven't generated a summary for this paper yet.