Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Natural Actor-Critic for Robust Reinforcement Learning with Function Approximation (2307.08875v2)

Published 17 Jul 2023 in cs.LG, cs.RO, and math.OC

Abstract: We study robust reinforcement learning (RL) with the goal of determining a well-performing policy that is robust against model mismatch between the training simulator and the testing environment. Previous policy-based robust RL algorithms mainly focus on the tabular setting under uncertainty sets that facilitate robust policy evaluation, but are no longer tractable when the number of states scales up. To this end, we propose two novel uncertainty set formulations, one based on double sampling and the other on an integral probability metric. Both make large-scale robust RL tractable even when one only has access to a simulator. We propose a robust natural actor-critic (RNAC) approach that incorporates the new uncertainty sets and employs function approximation. We provide finite-time convergence guarantees for the proposed RNAC algorithm to the optimal robust policy within the function approximation error. Finally, we demonstrate the robust performance of the policy learned by our proposed RNAC approach in multiple MuJoCo environments and a real-world TurtleBot navigation task.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Ruida Zhou (39 papers)
  2. Tao Liu (350 papers)
  3. Min Cheng (5 papers)
  4. Dileep Kalathil (62 papers)
  5. P. R. Kumar (78 papers)
  6. Chao Tian (78 papers)
Citations (13)

Summary

We haven't generated a summary for this paper yet.