Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A generalized stacked reinforcement learning method for sampled systems (2108.10392v3)

Published 23 Aug 2021 in cs.RO, cs.SY, eess.SY, and math.DS

Abstract: A common setting of reinforcement learning (RL) is a Markov decision process (MDP) in which the environment is a stochastic discrete-time dynamical system. Whereas MDPs are suitable in such applications as video-games or puzzles, physical systems are time-continuous. A general variant of RL is of digital format, where updates of the value (or cost) and policy are performed at discrete moments in time. The agent-environment loop then amounts to a sampled system, whereby sample-and-hold is a specific case. In this paper, we propose and benchmark two RL methods suitable for sampled systems. Specifically, we hybridize model-predictive control (MPC) with critics learning the optimal Q- and value (or cost-to-go) function. Optimality is analyzed and performance comparison is done in an experimental case study with a mobile robot.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Pavel Osinenko (35 papers)
  2. Dmitrii Dobriborsci (6 papers)
  3. Grigory Yaremenko (11 papers)
  4. Georgiy Malaniya (6 papers)
Citations (3)

Summary

We haven't generated a summary for this paper yet.