Learning Potential in Subgoal-Based Reward Shaping
Human knowledge can reduce the number of iterations required to learn in reinforcement learning. Though the most common approach uses trajectories, it is difficult to acquire them in certain domains. Subgoals, which are intermediate states, have been studied instead of trajectories. Subgoal-based re...
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2023-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/10047888/ |
_version_ | 1797894917594284032 |
---|---|
author | Takato Okudo Seiji Yamada |
author_facet | Takato Okudo Seiji Yamada |
author_sort | Takato Okudo |
collection | DOAJ |
description | Human knowledge can reduce the number of iterations required to learn in reinforcement learning. Though the most common approach uses trajectories, it is difficult to acquire them in certain domains. Subgoals, which are intermediate states, have been studied instead of trajectories. Subgoal-based reward shaping is a method that adds rewards to environmental rewards with a sequence of subgoals. The potential function, which is a component of subgoal-based reward shaping, is shaped by a hyperparameter that controls its output. However, it is not easy to select a hyperparameter because its appropriate value depends on the reward function of an environment, and the reward function is unknown but its output is available. We propose learned potential that parameterizes a hyperparameter and acquires its potential through learning. A value is an expected accumulated reward if an agent follows its policy after the current state and is strongly related to the reward function. With learned potential, we build an abstract state space, which is a higher-level representation of the state, with a sequence of subgoals and use the value over the abstract states as the potential to accelerate the value learning. N-step temporal-difference (TD) method learns the values over the abstract state. We conducted experiments to evaluate the effectiveness of learned potential, and the results indicate its effectiveness compared with a baseline reinforcement learning algorithm and several reward-shaping algorithms. The results also indicate that the participants’ subgoals are superior to subgoals generated randomly with learned potential. We discuss the appropriate number of subgoals for learned potential, that partially ordered subgoal is helpful for learned potential, that learned potential cannot make learning efficient in step penalized rewards, and that learned potential is superior to the non-learned potential in mixed positive and negative rewards. |
first_indexed | 2024-04-10T07:18:09Z |
format | Article |
id | doaj.art-188bcae17839430abf6a2e41c3bf08b6 |
institution | Directory Open Access Journal |
issn | 2169-3536 |
language | English |
last_indexed | 2024-04-10T07:18:09Z |
publishDate | 2023-01-01 |
publisher | IEEE |
record_format | Article |
series | IEEE Access |
spelling | doaj.art-188bcae17839430abf6a2e41c3bf08b62023-02-25T00:02:04ZengIEEEIEEE Access2169-35362023-01-0111171161713710.1109/ACCESS.2023.324626710047888Learning Potential in Subgoal-Based Reward ShapingTakato Okudo0https://orcid.org/0000-0002-7218-7842Seiji Yamada1https://orcid.org/0000-0002-5907-7382Department of Informatics, The Graduate University for Advanced Studies (SOKENDAI), Tokyo, JapanDepartment of Informatics, The Graduate University for Advanced Studies (SOKENDAI), Tokyo, JapanHuman knowledge can reduce the number of iterations required to learn in reinforcement learning. Though the most common approach uses trajectories, it is difficult to acquire them in certain domains. Subgoals, which are intermediate states, have been studied instead of trajectories. Subgoal-based reward shaping is a method that adds rewards to environmental rewards with a sequence of subgoals. The potential function, which is a component of subgoal-based reward shaping, is shaped by a hyperparameter that controls its output. However, it is not easy to select a hyperparameter because its appropriate value depends on the reward function of an environment, and the reward function is unknown but its output is available. We propose learned potential that parameterizes a hyperparameter and acquires its potential through learning. A value is an expected accumulated reward if an agent follows its policy after the current state and is strongly related to the reward function. With learned potential, we build an abstract state space, which is a higher-level representation of the state, with a sequence of subgoals and use the value over the abstract states as the potential to accelerate the value learning. N-step temporal-difference (TD) method learns the values over the abstract state. We conducted experiments to evaluate the effectiveness of learned potential, and the results indicate its effectiveness compared with a baseline reinforcement learning algorithm and several reward-shaping algorithms. The results also indicate that the participants’ subgoals are superior to subgoals generated randomly with learned potential. We discuss the appropriate number of subgoals for learned potential, that partially ordered subgoal is helpful for learned potential, that learned potential cannot make learning efficient in step penalized rewards, and that learned potential is superior to the non-learned potential in mixed positive and negative rewards.https://ieeexplore.ieee.org/document/10047888/Reinforcement learningdeep reinforcement learningsubgoalsreward shapingpotential-based reward shapingsubgoal-based reward shaping |
spellingShingle | Takato Okudo Seiji Yamada Learning Potential in Subgoal-Based Reward Shaping IEEE Access Reinforcement learning deep reinforcement learning subgoals reward shaping potential-based reward shaping subgoal-based reward shaping |
title | Learning Potential in Subgoal-Based Reward Shaping |
title_full | Learning Potential in Subgoal-Based Reward Shaping |
title_fullStr | Learning Potential in Subgoal-Based Reward Shaping |
title_full_unstemmed | Learning Potential in Subgoal-Based Reward Shaping |
title_short | Learning Potential in Subgoal-Based Reward Shaping |
title_sort | learning potential in subgoal based reward shaping |
topic | Reinforcement learning deep reinforcement learning subgoals reward shaping potential-based reward shaping subgoal-based reward shaping |
url | https://ieeexplore.ieee.org/document/10047888/ |
work_keys_str_mv | AT takatookudo learningpotentialinsubgoalbasedrewardshaping AT seijiyamada learningpotentialinsubgoalbasedrewardshaping |