Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
184 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Supplemental Material For "Primal-Dual Q-Learning Framework for LQR Design" (1811.08475v1)

Published 20 Nov 2018 in math.OC

Abstract: Recently, reinforcement learning (RL) is receiving more and more attentions due to its successful demonstrations outperforming human performance in certain challenging tasks. In our paper `primal-dual Q-learning framework for LQR design,' we study a new optimization formulation of the linear quadratic regulator (LQR) problem via the Lagrangian duality theories in order to lay theoretical foundations of potentially effective RL algorithms. The new optimization problem includes the Q-function parameters so that it can be directly used to develop Q-learning algorithms, known to be one of the most popular RL algorithms. In the paper, we prove relations between saddle-points of the Lagrangian function and the optimal solutions of the BeLLMan equation. As an application, we propose a model-free primal-dual Q-learning algorithm to solve the LQR problem and demonstrate its validity through examples. It is meaningful to consider additional potential applications of the proposed analysis. Various SDP formulations of Problem 5 or Problem 2 of the paper can be derived, and they can be used to develop new analysis and control design approaches. For example, an SDP-based optimal control design with energy and input constraints can be derived. Another direction is algorithms for structured controller designs. These approaches are included in this supplemental material.

Summary

We haven't generated a summary for this paper yet.