Model Predictive Control-Based Reinforcement Learning Using Expected Sarsa

Recent studies have shown the potential of Reinforcement Learning (RL) algorithms in tuning the parameters of Model Predictive Controllers (MPC), including the weights of the cost function and unknown parameters of the MPC model. However, a framework for easy and straightforward implementation that...

Full description

Bibliographic Details
Main Authors: Hoomaan Moradimaryamnegari, Marco Frego, Angelika Peer
Format: Article
Language:English
Published: IEEE 2022-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/9846979/
_version_ 1811342517888614400
author Hoomaan Moradimaryamnegari
Marco Frego
Angelika Peer
author_facet Hoomaan Moradimaryamnegari
Marco Frego
Angelika Peer
author_sort Hoomaan Moradimaryamnegari
collection DOAJ
description Recent studies have shown the potential of Reinforcement Learning (RL) algorithms in tuning the parameters of Model Predictive Controllers (MPC), including the weights of the cost function and unknown parameters of the MPC model. However, a framework for easy and straightforward implementation that allows training in just a few episodes and overcoming the need for imposing extra constraints as required by state-of-the-art methods, is still missing. In this study, we present two implementations to achieve these goals. In the first approach, a nonlinear MPC plays the role of a function approximator for an Expected Sarsa RL algorithm. In the second approach, only the MPC cost function is considered as the function approximator, while the unknown parameters of the MPC model are updated based on more classical system identification. In order to evaluate the performance of the proposed algorithms, first numerical simulations are performed on a coupled tanks system. Then, both algorithms are applied to the real system and their closed-loop performance and convergence speed are compared with each other. The results indicate that the proposed algorithms allow tuning of MPCs over very few episodes. Finally, also the disturbance rejection ability of the proposed methods is demonstrated.
first_indexed 2024-04-13T19:12:25Z
format Article
id doaj.art-80328c61febd45e0a03c6025a35763c6
institution Directory Open Access Journal
issn 2169-3536
language English
last_indexed 2024-04-13T19:12:25Z
publishDate 2022-01-01
publisher IEEE
record_format Article
series IEEE Access
spelling doaj.art-80328c61febd45e0a03c6025a35763c62022-12-22T02:33:46ZengIEEEIEEE Access2169-35362022-01-0110811778119110.1109/ACCESS.2022.31955309846979Model Predictive Control-Based Reinforcement Learning Using Expected SarsaHoomaan Moradimaryamnegari0https://orcid.org/0000-0002-0858-0420Marco Frego1https://orcid.org/0000-0003-2855-9052Angelika Peer2https://orcid.org/0000-0002-2896-9011Human-Centered Technologies and Machine Intelligence Laboratory, Faculty of Science and Technology, Free University of Bozen-Bolzano, Bolzano, ItalyHuman-Centered Technologies and Machine Intelligence Laboratory, Faculty of Science and Technology, Free University of Bozen-Bolzano, Bolzano, ItalyHuman-Centered Technologies and Machine Intelligence Laboratory, Faculty of Science and Technology, Free University of Bozen-Bolzano, Bolzano, ItalyRecent studies have shown the potential of Reinforcement Learning (RL) algorithms in tuning the parameters of Model Predictive Controllers (MPC), including the weights of the cost function and unknown parameters of the MPC model. However, a framework for easy and straightforward implementation that allows training in just a few episodes and overcoming the need for imposing extra constraints as required by state-of-the-art methods, is still missing. In this study, we present two implementations to achieve these goals. In the first approach, a nonlinear MPC plays the role of a function approximator for an Expected Sarsa RL algorithm. In the second approach, only the MPC cost function is considered as the function approximator, while the unknown parameters of the MPC model are updated based on more classical system identification. In order to evaluate the performance of the proposed algorithms, first numerical simulations are performed on a coupled tanks system. Then, both algorithms are applied to the real system and their closed-loop performance and convergence speed are compared with each other. The results indicate that the proposed algorithms allow tuning of MPCs over very few episodes. Finally, also the disturbance rejection ability of the proposed methods is demonstrated.https://ieeexplore.ieee.org/document/9846979/Model predictive controlreinforcement learningexpected Sarsa algorithmmodel-based learning method
spellingShingle Hoomaan Moradimaryamnegari
Marco Frego
Angelika Peer
Model Predictive Control-Based Reinforcement Learning Using Expected Sarsa
IEEE Access
Model predictive control
reinforcement learning
expected Sarsa algorithm
model-based learning method
title Model Predictive Control-Based Reinforcement Learning Using Expected Sarsa
title_full Model Predictive Control-Based Reinforcement Learning Using Expected Sarsa
title_fullStr Model Predictive Control-Based Reinforcement Learning Using Expected Sarsa
title_full_unstemmed Model Predictive Control-Based Reinforcement Learning Using Expected Sarsa
title_short Model Predictive Control-Based Reinforcement Learning Using Expected Sarsa
title_sort model predictive control based reinforcement learning using expected sarsa
topic Model predictive control
reinforcement learning
expected Sarsa algorithm
model-based learning method
url https://ieeexplore.ieee.org/document/9846979/
work_keys_str_mv AT hoomaanmoradimaryamnegari modelpredictivecontrolbasedreinforcementlearningusingexpectedsarsa
AT marcofrego modelpredictivecontrolbasedreinforcementlearningusingexpectedsarsa
AT angelikapeer modelpredictivecontrolbasedreinforcementlearningusingexpectedsarsa