Efficient Elitist Cooperative Evolutionary Algorithm for Multi-Objective Reinforcement Learning

Sequential decision-making problems with multiple objectives are known as multi-objective reinforcement learning. In these scenarios, decision-makers require a complete Pareto front that consists of Pareto optimal solutions. Such a front enables decision-makers to understand the relationship between...

Full description

Bibliographic Details
Main Authors: Dan Zhou, Jiqing Du, Sachiyo Arai
Format: Article
Language:English
Published: IEEE 2023-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/10113623/
_version_ 1797829122699821056
author Dan Zhou
Jiqing Du
Sachiyo Arai
author_facet Dan Zhou
Jiqing Du
Sachiyo Arai
author_sort Dan Zhou
collection DOAJ
description Sequential decision-making problems with multiple objectives are known as multi-objective reinforcement learning. In these scenarios, decision-makers require a complete Pareto front that consists of Pareto optimal solutions. Such a front enables decision-makers to understand the relationship between objectives and make informed decisions from a broad range of solutions. However, existing methods may be unable to search for solutions in concave regions of the Pareto front or lack global optimization ability, leading to incomplete Pareto fronts. To address this issue, we propose an efficient elitist cooperative evolutionary algorithm that maintains both an evolving population and an elite archive. The elite archive uses cooperative operations with various genetic operators to guide the evolving population, resulting in efficient searches for Pareto optimal solutions. The experimental results on submarine treasure hunting benchmarks demonstrate the effectiveness of the proposed method in solving various multi-objective reinforcement learning problems and providing decision-makers with a set of trade-off solutions between travel time and treasure amount, enabling them to make flexible and informed decisions based on their preferences. Therefore, the proposed method has the potential to be a useful tool for implementing real-world applications.
first_indexed 2024-04-09T13:15:24Z
format Article
id doaj.art-3c2dcab280274e4f994b6f59af3e4f83
institution Directory Open Access Journal
issn 2169-3536
language English
last_indexed 2024-04-09T13:15:24Z
publishDate 2023-01-01
publisher IEEE
record_format Article
series IEEE Access
spelling doaj.art-3c2dcab280274e4f994b6f59af3e4f832023-05-11T23:00:47ZengIEEEIEEE Access2169-35362023-01-0111431284313910.1109/ACCESS.2023.327211510113623Efficient Elitist Cooperative Evolutionary Algorithm for Multi-Objective Reinforcement LearningDan Zhou0https://orcid.org/0000-0002-4875-1781Jiqing Du1Sachiyo Arai2Department of Urban Environment Systems, Division of Earth and Environmental Sciences, Graduate School of Science and Engineering, Chiba University, Chiba, JapanDepartment of Urban Environment Systems, Division of Earth and Environmental Sciences, Graduate School of Science and Engineering, Chiba University, Chiba, JapanDepartment of Urban Environment Systems, Division of Earth and Environmental Sciences, Graduate School of Science and Engineering, Chiba University, Chiba, JapanSequential decision-making problems with multiple objectives are known as multi-objective reinforcement learning. In these scenarios, decision-makers require a complete Pareto front that consists of Pareto optimal solutions. Such a front enables decision-makers to understand the relationship between objectives and make informed decisions from a broad range of solutions. However, existing methods may be unable to search for solutions in concave regions of the Pareto front or lack global optimization ability, leading to incomplete Pareto fronts. To address this issue, we propose an efficient elitist cooperative evolutionary algorithm that maintains both an evolving population and an elite archive. The elite archive uses cooperative operations with various genetic operators to guide the evolving population, resulting in efficient searches for Pareto optimal solutions. The experimental results on submarine treasure hunting benchmarks demonstrate the effectiveness of the proposed method in solving various multi-objective reinforcement learning problems and providing decision-makers with a set of trade-off solutions between travel time and treasure amount, enabling them to make flexible and informed decisions based on their preferences. Therefore, the proposed method has the potential to be a useful tool for implementing real-world applications.https://ieeexplore.ieee.org/document/10113623/Multi-objective reinforcement learningefficientcooperativePareto frontelite archive
spellingShingle Dan Zhou
Jiqing Du
Sachiyo Arai
Efficient Elitist Cooperative Evolutionary Algorithm for Multi-Objective Reinforcement Learning
IEEE Access
Multi-objective reinforcement learning
efficient
cooperative
Pareto front
elite archive
title Efficient Elitist Cooperative Evolutionary Algorithm for Multi-Objective Reinforcement Learning
title_full Efficient Elitist Cooperative Evolutionary Algorithm for Multi-Objective Reinforcement Learning
title_fullStr Efficient Elitist Cooperative Evolutionary Algorithm for Multi-Objective Reinforcement Learning
title_full_unstemmed Efficient Elitist Cooperative Evolutionary Algorithm for Multi-Objective Reinforcement Learning
title_short Efficient Elitist Cooperative Evolutionary Algorithm for Multi-Objective Reinforcement Learning
title_sort efficient elitist cooperative evolutionary algorithm for multi objective reinforcement learning
topic Multi-objective reinforcement learning
efficient
cooperative
Pareto front
elite archive
url https://ieeexplore.ieee.org/document/10113623/
work_keys_str_mv AT danzhou efficientelitistcooperativeevolutionaryalgorithmformultiobjectivereinforcementlearning
AT jiqingdu efficientelitistcooperativeevolutionaryalgorithmformultiobjectivereinforcementlearning
AT sachiyoarai efficientelitistcooperativeevolutionaryalgorithmformultiobjectivereinforcementlearning