Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Tractable Algorithm For Finite-Horizon Continuous Reinforcement Learning (1906.11245v1)

Published 26 Jun 2019 in cs.LG and cs.AI

Abstract: We consider the finite horizon continuous reinforcement learning problem. Our contribution is three-fold. First,we give a tractable algorithm based on optimistic value iteration for the problem. Next,we give a lower bound on regret of order $\Omega(T{2/3})$ for any algorithm discretizes the state space, improving the previous regret bound of $\Omega(T{1/2})$ of Ortner and Ryabko \cite{contrl} for the same problem. Next,under the assumption that the rewards and transitions are H\"{o}lder Continuous we show that the upper bound on the discretization error is $const.Ln{-\alpha}T$. Finally,we give some simple experiments to validate our propositions.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
Citations (1)

Summary

We haven't generated a summary for this paper yet.