Comparison of Reinforcement Learning and Model Predictive Control for Automated Generation of Optimal Control for Dynamic Systems within a Design Space Exploration Framework

This work provides a study of methods for the automated derivation of control strategies for over-actuated systems. For this purpose, Reinforcement Learning (RL) and Model Predictive Control (MPC) approximating the solution of the Optimal Control Problem (OCP) are compared using the example of an ov...

Full description

Bibliographic Details
Main Authors: Patrick Hoffmann, Kirill Gorelik, Valentin Ivanov
Format: Article
Language:English
Published: Society of Automotive Engineers of Japan, Inc. 2024-01-01
Series:International Journal of Automotive Engineering
Online Access:https://www.jstage.jst.go.jp/article/jsaeijae/15/1/15_20244099/_article/-char/ja
_version_ 1797272359350042624
author Patrick Hoffmann
Kirill Gorelik
Valentin Ivanov
author_facet Patrick Hoffmann
Kirill Gorelik
Valentin Ivanov
author_sort Patrick Hoffmann
collection DOAJ
description This work provides a study of methods for the automated derivation of control strategies for over-actuated systems. For this purpose, Reinforcement Learning (RL) and Model Predictive Control (MPC) approximating the solution of the Optimal Control Problem (OCP) are compared using the example of an over-actuated vehicle model executing an ISO Double Lane Change (DLC). This exemplary driving maneuver is chosen due to its critical vehicle dynamics for the comparison of algorithms in terms of control performance and possible automation within a design space exploration framework. The algorithms show reasonable control results for the goal of this study, although there are differences in terms of driving stability. While Model Predictive Control first requires the optimization of the trajectory, which should then be optimally tracked, RL may combine both in one step. In addition, manual effort required to adapt the OCP problem to new design variants for solving it with RL and MPC is evaluated and assessed with respect to its automation. As a result of this study, an Actor-Critic Reinforcement Learning method is recommended for the automated derivation of control strategies in the context of a design space exploration.
first_indexed 2024-03-07T14:27:11Z
format Article
id doaj.art-80123fe37d8047e6a33df25617201354
institution Directory Open Access Journal
issn 2185-0992
language English
last_indexed 2024-03-07T14:27:11Z
publishDate 2024-01-01
publisher Society of Automotive Engineers of Japan, Inc.
record_format Article
series International Journal of Automotive Engineering
spelling doaj.art-80123fe37d8047e6a33df256172013542024-03-06T07:44:10ZengSociety of Automotive Engineers of Japan, Inc.International Journal of Automotive Engineering2185-09922024-01-01151192610.20485/jsaeijae.15.1_19Comparison of Reinforcement Learning and Model Predictive Control for Automated Generation of Optimal Control for Dynamic Systems within a Design Space Exploration FrameworkPatrick Hoffmann0Kirill Gorelik1Valentin Ivanov2Robert Bosch GmbHRobert Bosch GmbHSmart Vehicle Systems - Working GroupThis work provides a study of methods for the automated derivation of control strategies for over-actuated systems. For this purpose, Reinforcement Learning (RL) and Model Predictive Control (MPC) approximating the solution of the Optimal Control Problem (OCP) are compared using the example of an over-actuated vehicle model executing an ISO Double Lane Change (DLC). This exemplary driving maneuver is chosen due to its critical vehicle dynamics for the comparison of algorithms in terms of control performance and possible automation within a design space exploration framework. The algorithms show reasonable control results for the goal of this study, although there are differences in terms of driving stability. While Model Predictive Control first requires the optimization of the trajectory, which should then be optimally tracked, RL may combine both in one step. In addition, manual effort required to adapt the OCP problem to new design variants for solving it with RL and MPC is evaluated and assessed with respect to its automation. As a result of this study, an Actor-Critic Reinforcement Learning method is recommended for the automated derivation of control strategies in the context of a design space exploration.https://www.jstage.jst.go.jp/article/jsaeijae/15/1/15_20244099/_article/-char/ja
spellingShingle Patrick Hoffmann
Kirill Gorelik
Valentin Ivanov
Comparison of Reinforcement Learning and Model Predictive Control for Automated Generation of Optimal Control for Dynamic Systems within a Design Space Exploration Framework
International Journal of Automotive Engineering
title Comparison of Reinforcement Learning and Model Predictive Control for Automated Generation of Optimal Control for Dynamic Systems within a Design Space Exploration Framework
title_full Comparison of Reinforcement Learning and Model Predictive Control for Automated Generation of Optimal Control for Dynamic Systems within a Design Space Exploration Framework
title_fullStr Comparison of Reinforcement Learning and Model Predictive Control for Automated Generation of Optimal Control for Dynamic Systems within a Design Space Exploration Framework
title_full_unstemmed Comparison of Reinforcement Learning and Model Predictive Control for Automated Generation of Optimal Control for Dynamic Systems within a Design Space Exploration Framework
title_short Comparison of Reinforcement Learning and Model Predictive Control for Automated Generation of Optimal Control for Dynamic Systems within a Design Space Exploration Framework
title_sort comparison of reinforcement learning and model predictive control for automated generation of optimal control for dynamic systems within a design space exploration framework
url https://www.jstage.jst.go.jp/article/jsaeijae/15/1/15_20244099/_article/-char/ja
work_keys_str_mv AT patrickhoffmann comparisonofreinforcementlearningandmodelpredictivecontrolforautomatedgenerationofoptimalcontrolfordynamicsystemswithinadesignspaceexplorationframework
AT kirillgorelik comparisonofreinforcementlearningandmodelpredictivecontrolforautomatedgenerationofoptimalcontrolfordynamicsystemswithinadesignspaceexplorationframework
AT valentinivanov comparisonofreinforcementlearningandmodelpredictivecontrolforautomatedgenerationofoptimalcontrolfordynamicsystemswithinadesignspaceexplorationframework