Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Adversarial Skill Chaining for Long-Horizon Robot Manipulation via Terminal State Regularization (2111.07999v1)

Published 15 Nov 2021 in cs.LG, cs.AI, and cs.RO

Abstract: Skill chaining is a promising approach for synthesizing complex behaviors by sequentially combining previously learned skills. Yet, a naive composition of skills fails when a policy encounters a starting state never seen during its training. For successful skill chaining, prior approaches attempt to widen the policy's starting state distribution. However, these approaches require larger state distributions to be covered as more policies are sequenced, and thus are limited to short skill sequences. In this paper, we propose to chain multiple policies without excessively large initial state distributions by regularizing the terminal state distributions in an adversarial learning framework. We evaluate our approach on two complex long-horizon manipulation tasks of furniture assembly. Our results have shown that our method establishes the first model-free reinforcement learning algorithm to solve these tasks; whereas prior skill chaining approaches fail. The code and videos are available at https://clvrai.com/skill-chaining

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Youngwoon Lee (23 papers)
  2. Joseph J. Lim (36 papers)
  3. Anima Anandkumar (236 papers)
  4. Yuke Zhu (134 papers)
Citations (34)

Summary

We haven't generated a summary for this paper yet.