Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Vision-based Robot Manipulation Learning via Human Demonstrations (2003.00385v1)

Published 1 Mar 2020 in cs.RO

Abstract: Vision-based learning methods provide promise for robots to learn complex manipulation tasks. However, how to generalize the learned manipulation skills to real-world interactions remains an open question. In this work, we study robotic manipulation skill learning from a single third-person view demonstration by using activity recognition and object detection in computer vision. To facilitate generalization across objects and environments, we propose to use a prior knowledge base in the form of a text corpus to infer the object to be interacted with in the context of a robot. We evaluate our approach in a real-world robot, using several simple and complex manipulation tasks commonly performed in daily life. The experimental results show that our approach achieves good generalization performance even from small amounts of training data.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Zhixin Jia (2 papers)
  2. Mengxiang Lin (8 papers)
  3. Zhixin Chen (10 papers)
  4. Shibo Jian (2 papers)
Citations (7)

Summary

We haven't generated a summary for this paper yet.