A Unifying Framework for Reinforcement Learning and Planning

Sequential decision making, commonly formalized as optimization of a Markov Decision Process, is a key challenge in artificial intelligence. Two successful approaches to MDP optimization are reinforcement learning and planning, which both largely have their own research communities. However, if both...

Full description

Bibliographic Details
Main Authors: Thomas M. Moerland, Joost Broekens, Aske Plaat, Catholijn M. Jonker
Format: Article
Language:English
Published: Frontiers Media S.A. 2022-07-01
Series:Frontiers in Artificial Intelligence
Subjects:
Online Access:https://www.frontiersin.org/articles/10.3389/frai.2022.908353/full
_version_ 1818518958479769600
author Thomas M. Moerland
Joost Broekens
Aske Plaat
Catholijn M. Jonker
Catholijn M. Jonker
author_facet Thomas M. Moerland
Joost Broekens
Aske Plaat
Catholijn M. Jonker
Catholijn M. Jonker
author_sort Thomas M. Moerland
collection DOAJ
description Sequential decision making, commonly formalized as optimization of a Markov Decision Process, is a key challenge in artificial intelligence. Two successful approaches to MDP optimization are reinforcement learning and planning, which both largely have their own research communities. However, if both research fields solve the same problem, then we might be able to disentangle the common factors in their solution approaches. Therefore, this paper presents a unifying algorithmic framework for reinforcement learning and planning (FRAP), which identifies underlying dimensions on which MDP planning and learning algorithms have to decide. At the end of the paper, we compare a variety of well-known planning, model-free and model-based RL algorithms along these dimensions. Altogether, the framework may help provide deeper insight in the algorithmic design space of planning and reinforcement learning.
first_indexed 2024-12-11T01:17:21Z
format Article
id doaj.art-1964865081014a27b68c9d4d7dda37c8
institution Directory Open Access Journal
issn 2624-8212
language English
last_indexed 2024-12-11T01:17:21Z
publishDate 2022-07-01
publisher Frontiers Media S.A.
record_format Article
series Frontiers in Artificial Intelligence
spelling doaj.art-1964865081014a27b68c9d4d7dda37c82022-12-22T01:25:50ZengFrontiers Media S.A.Frontiers in Artificial Intelligence2624-82122022-07-01510.3389/frai.2022.908353908353A Unifying Framework for Reinforcement Learning and PlanningThomas M. Moerland0Joost Broekens1Aske Plaat2Catholijn M. Jonker3Catholijn M. Jonker4Leiden Institute of Advanced Computer Science (LIACS), Leiden University, Leiden, NetherlandsLeiden Institute of Advanced Computer Science (LIACS), Leiden University, Leiden, NetherlandsLeiden Institute of Advanced Computer Science (LIACS), Leiden University, Leiden, NetherlandsLeiden Institute of Advanced Computer Science (LIACS), Leiden University, Leiden, NetherlandsInteractive Intelligence, Delft University of Technology, Delft, NetherlandsSequential decision making, commonly formalized as optimization of a Markov Decision Process, is a key challenge in artificial intelligence. Two successful approaches to MDP optimization are reinforcement learning and planning, which both largely have their own research communities. However, if both research fields solve the same problem, then we might be able to disentangle the common factors in their solution approaches. Therefore, this paper presents a unifying algorithmic framework for reinforcement learning and planning (FRAP), which identifies underlying dimensions on which MDP planning and learning algorithms have to decide. At the end of the paper, we compare a variety of well-known planning, model-free and model-based RL algorithms along these dimensions. Altogether, the framework may help provide deeper insight in the algorithmic design space of planning and reinforcement learning.https://www.frontiersin.org/articles/10.3389/frai.2022.908353/fullplanningreinforcement learningmodel-based reinforcement learningframeworkoverviewsynthesis
spellingShingle Thomas M. Moerland
Joost Broekens
Aske Plaat
Catholijn M. Jonker
Catholijn M. Jonker
A Unifying Framework for Reinforcement Learning and Planning
Frontiers in Artificial Intelligence
planning
reinforcement learning
model-based reinforcement learning
framework
overview
synthesis
title A Unifying Framework for Reinforcement Learning and Planning
title_full A Unifying Framework for Reinforcement Learning and Planning
title_fullStr A Unifying Framework for Reinforcement Learning and Planning
title_full_unstemmed A Unifying Framework for Reinforcement Learning and Planning
title_short A Unifying Framework for Reinforcement Learning and Planning
title_sort unifying framework for reinforcement learning and planning
topic planning
reinforcement learning
model-based reinforcement learning
framework
overview
synthesis
url https://www.frontiersin.org/articles/10.3389/frai.2022.908353/full
work_keys_str_mv AT thomasmmoerland aunifyingframeworkforreinforcementlearningandplanning
AT joostbroekens aunifyingframeworkforreinforcementlearningandplanning
AT askeplaat aunifyingframeworkforreinforcementlearningandplanning
AT catholijnmjonker aunifyingframeworkforreinforcementlearningandplanning
AT catholijnmjonker aunifyingframeworkforreinforcementlearningandplanning
AT thomasmmoerland unifyingframeworkforreinforcementlearningandplanning
AT joostbroekens unifyingframeworkforreinforcementlearningandplanning
AT askeplaat unifyingframeworkforreinforcementlearningandplanning
AT catholijnmjonker unifyingframeworkforreinforcementlearningandplanning
AT catholijnmjonker unifyingframeworkforreinforcementlearningandplanning