Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

State Augmented Constrained Reinforcement Learning: Overcoming the Limitations of Learning with Rewards (2102.11941v2)

Published 23 Feb 2021 in cs.LG, cs.RO, and math.OC

Abstract: A common formulation of constrained reinforcement learning involves multiple rewards that must individually accumulate to given thresholds. In this class of problems, we show a simple example in which the desired optimal policy cannot be induced by any weighted linear combination of rewards. Hence, there exist constrained reinforcement learning problems for which neither regularized nor classical primal-dual methods yield optimal policies. This work addresses this shortcoming by augmenting the state with Lagrange multipliers and reinterpreting primal-dual methods as the portion of the dynamics that drives the multipliers evolution. This approach provides a systematic state augmentation procedure that is guaranteed to solve reinforcement learning problems with constraints. Thus, as we illustrate by an example, while previous methods can fail at finding optimal policies, running the dual dynamics while executing the augmented policy yields an algorithm that provably samples actions from the optimal policy.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Miguel Calvo-Fullana (21 papers)
  2. Santiago Paternain (50 papers)
  3. Luiz F. O. Chamon (38 papers)
  4. Alejandro Ribeiro (281 papers)
Citations (22)

Summary

We haven't generated a summary for this paper yet.