Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

On the Global Convergence of Imitation Learning: A Case for Linear Quadratic Regulator (1901.03674v1)

Published 11 Jan 2019 in cs.LG, cs.AI, math.OC, and stat.ML

Abstract: We study the global convergence of generative adversarial imitation learning for linear quadratic regulators, which is posed as minimax optimization. To address the challenges arising from non-convex-concave geometry, we analyze the alternating gradient algorithm and establish its Q-linear rate of convergence to a unique saddle point, which simultaneously recovers the globally optimal policy and reward function. We hope our results may serve as a small step towards understanding and taming the instability in imitation learning as well as in more general non-convex-concave alternating minimax optimization that arises from reinforcement learning and generative adversarial learning.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Qi Cai (40 papers)
  2. Mingyi Hong (172 papers)
  3. Yongxin Chen (146 papers)
  4. Zhaoran Wang (164 papers)
Citations (33)