Battery Energy Management in a Microgrid Using Batch Reinforcement Learning

Motivated by recent developments in batch Reinforcement Learning (RL), this paper contributes to the application of batch RL in energy management in microgrids. We tackle the challenge of finding a closed-loop control policy to optimally schedule the operation of a storage device, in order to maximi...

Full description

Bibliographic Details
Main Authors: Brida V. Mbuwir, Frederik Ruelens, Fred Spiessens, Geert Deconinck
Format: Article
Language:English
Published: MDPI AG 2017-11-01
Series:Energies
Subjects:
Online Access:https://www.mdpi.com/1996-1073/10/11/1846
_version_ 1818014136997511168
author Brida V. Mbuwir
Frederik Ruelens
Fred Spiessens
Geert Deconinck
author_facet Brida V. Mbuwir
Frederik Ruelens
Fred Spiessens
Geert Deconinck
author_sort Brida V. Mbuwir
collection DOAJ
description Motivated by recent developments in batch Reinforcement Learning (RL), this paper contributes to the application of batch RL in energy management in microgrids. We tackle the challenge of finding a closed-loop control policy to optimally schedule the operation of a storage device, in order to maximize self-consumption of local photovoltaic production in a microgrid. In this work, the fitted Q-iteration algorithm, a standard batch RL technique, is used by an RL agent to construct a control policy. The proposed method is data-driven and uses a state-action value function to find an optimal scheduling plan for a battery. The battery’s charge and discharge efficiencies, and the nonlinearity in the microgrid due to the inverter’s efficiency are taken into account. The proposed approach has been tested by simulation in a residential setting using data from Belgian residential consumers. The developed framework is benchmarked with a model-based technique, and the simulation results show a performance gap of 19%. The simulation results provide insight for developing optimal policies in more realistically-scaled and interconnected microgrids and for including uncertainties in generation and consumption for which white-box models become inaccurate and/or infeasible.
first_indexed 2024-04-14T06:41:17Z
format Article
id doaj.art-b11ff664232841f586419359d580f100
institution Directory Open Access Journal
issn 1996-1073
language English
last_indexed 2024-04-14T06:41:17Z
publishDate 2017-11-01
publisher MDPI AG
record_format Article
series Energies
spelling doaj.art-b11ff664232841f586419359d580f1002022-12-22T02:07:19ZengMDPI AGEnergies1996-10732017-11-011011184610.3390/en10111846en10111846Battery Energy Management in a Microgrid Using Batch Reinforcement LearningBrida V. Mbuwir0Frederik Ruelens1Fred Spiessens2Geert Deconinck3ESAT/Electa, KU Leuven, Kasteelpark Arenberg 10 bus 2445, BE-3001 Leuven, BelgiumESAT/Electa, KU Leuven, Kasteelpark Arenberg 10 bus 2445, BE-3001 Leuven, BelgiumEnergy Department, EnergyVille, Thor Park, Poort Genk 8130, 3600 Genk, BelgiumESAT/Electa, KU Leuven, Kasteelpark Arenberg 10 bus 2445, BE-3001 Leuven, BelgiumMotivated by recent developments in batch Reinforcement Learning (RL), this paper contributes to the application of batch RL in energy management in microgrids. We tackle the challenge of finding a closed-loop control policy to optimally schedule the operation of a storage device, in order to maximize self-consumption of local photovoltaic production in a microgrid. In this work, the fitted Q-iteration algorithm, a standard batch RL technique, is used by an RL agent to construct a control policy. The proposed method is data-driven and uses a state-action value function to find an optimal scheduling plan for a battery. The battery’s charge and discharge efficiencies, and the nonlinearity in the microgrid due to the inverter’s efficiency are taken into account. The proposed approach has been tested by simulation in a residential setting using data from Belgian residential consumers. The developed framework is benchmarked with a model-based technique, and the simulation results show a performance gap of 19%. The simulation results provide insight for developing optimal policies in more realistically-scaled and interconnected microgrids and for including uncertainties in generation and consumption for which white-box models become inaccurate and/or infeasible.https://www.mdpi.com/1996-1073/10/11/1846control policyfitted-Q iterationmicrogridsreinforcement learning
spellingShingle Brida V. Mbuwir
Frederik Ruelens
Fred Spiessens
Geert Deconinck
Battery Energy Management in a Microgrid Using Batch Reinforcement Learning
Energies
control policy
fitted-Q iteration
microgrids
reinforcement learning
title Battery Energy Management in a Microgrid Using Batch Reinforcement Learning
title_full Battery Energy Management in a Microgrid Using Batch Reinforcement Learning
title_fullStr Battery Energy Management in a Microgrid Using Batch Reinforcement Learning
title_full_unstemmed Battery Energy Management in a Microgrid Using Batch Reinforcement Learning
title_short Battery Energy Management in a Microgrid Using Batch Reinforcement Learning
title_sort battery energy management in a microgrid using batch reinforcement learning
topic control policy
fitted-Q iteration
microgrids
reinforcement learning
url https://www.mdpi.com/1996-1073/10/11/1846
work_keys_str_mv AT bridavmbuwir batteryenergymanagementinamicrogridusingbatchreinforcementlearning
AT frederikruelens batteryenergymanagementinamicrogridusingbatchreinforcementlearning
AT fredspiessens batteryenergymanagementinamicrogridusingbatchreinforcementlearning
AT geertdeconinck batteryenergymanagementinamicrogridusingbatchreinforcementlearning