Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Unifying Framework for Reinforcement Learning and Planning (2006.15009v4)

Published 26 Jun 2020 in cs.LG, cs.AI, cs.RO, and stat.ML

Abstract: Sequential decision making, commonly formalized as optimization of a Markov Decision Process, is a key challenge in artificial intelligence. Two successful approaches to MDP optimization are reinforcement learning and planning, which both largely have their own research communities. However, if both research fields solve the same problem, then we might be able to disentangle the common factors in their solution approaches. Therefore, this paper presents a unifying algorithmic framework for reinforcement learning and planning (FRAP), which identifies underlying dimensions on which MDP planning and learning algorithms have to decide. At the end of the paper, we compare a variety of well-known planning, model-free and model-based RL algorithms along these dimensions. Altogether, the framework may help provide deeper insight in the algorithmic design space of planning and reinforcement learning.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Thomas M. Moerland (24 papers)
  2. Joost Broekens (22 papers)
  3. Aske Plaat (76 papers)
  4. Catholijn M. Jonker (42 papers)
Citations (8)

Summary

We haven't generated a summary for this paper yet.