Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

AW-Opt: Learning Robotic Skills with Imitation and Reinforcement at Scale (2111.05424v2)

Published 9 Nov 2021 in cs.RO

Abstract: Robotic skills can be learned via imitation learning (IL) using user-provided demonstrations, or via reinforcement learning (RL) using large amountsof autonomously collected experience.Both methods have complementarystrengths and weaknesses: RL can reach a high level of performance, but requiresexploration, which can be very time consuming and unsafe; IL does not requireexploration, but only learns skills that are as good as the provided demonstrations.Can a single method combine the strengths of both approaches? A number ofprior methods have aimed to address this question, proposing a variety of tech-niques that integrate elements of IL and RL. However, scaling up such methodsto complex robotic skills that integrate diverse offline data and generalize mean-ingfully to real-world scenarios still presents a major challenge. In this paper, ouraim is to test the scalability of prior IL + RL algorithms and devise a system basedon detailed empirical experimentation that combines existing components in themost effective and scalable way. To that end, we present a series of experimentsaimed at understanding the implications of each design decision, so as to develop acombined approach that can utilize demonstrations and heterogeneous prior datato attain the best performance on a range of real-world and realistic simulatedrobotic problems. Our complete method, which we call AW-Opt, combines ele-ments of advantage-weighted regression [1, 2] and QT-Opt [3], providing a unifiedapproach for integrating demonstrations and offline data for robotic manipulation.Please see https://awopt.github.io for more details.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (11)
  1. Yao Lu (212 papers)
  2. Karol Hausman (56 papers)
  3. Yevgen Chebotar (28 papers)
  4. Mengyuan Yan (14 papers)
  5. Eric Jang (19 papers)
  6. Alexander Herzog (32 papers)
  7. Ted Xiao (40 papers)
  8. Alex Irpan (23 papers)
  9. Mohi Khansari (18 papers)
  10. Dmitry Kalashnikov (34 papers)
  11. Sergey Levine (531 papers)
Citations (57)
Github Logo Streamline Icon: https://streamlinehq.com

GitHub