Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

State-Conditioned Adversarial Subgoal Generation (2201.09635v4)

Published 24 Jan 2022 in cs.LG

Abstract: Hierarchical reinforcement learning (HRL) proposes to solve difficult tasks by performing decision-making and control at successively higher levels of temporal abstraction. However, off-policy HRL often suffers from the problem of a non-stationary high-level policy since the low-level policy is constantly changing. In this paper, we propose a novel HRL approach for mitigating the non-stationarity by adversarially enforcing the high-level policy to generate subgoals compatible with the current instantiation of the low-level policy. In practice, the adversarial learning is implemented by training a simple state-conditioned discriminator network concurrently with the high-level policy which determines the compatibility level of subgoals. Comparison to state-of-the-art algorithms shows that our approach improves both learning efficiency and performance in challenging continuous control tasks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Vivienne Huiling Wang (3 papers)
  2. Joni Pajarinen (68 papers)
  3. Tinghuai Wang (12 papers)
  4. Joni-Kristian Kämäräinen (32 papers)
Citations (8)

Summary

We haven't generated a summary for this paper yet.