Learning Potential in Subgoal-Based Reward Shaping
Human knowledge can reduce the number of iterations required to learn in reinforcement learning. Though the most common approach uses trajectories, it is difficult to acquire them in certain domains. Subgoals, which are intermediate states, have been studied instead of trajectories. Subgoal-based re...
Main Authors: | Takato Okudo, Seiji Yamada |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2023-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/10047888/ |
Similar Items
-
Reward Shaping Based Federated Reinforcement Learning
by: Yiqiu Hu, et al.
Published: (2021-01-01) -
Rewards Prediction-Based Credit Assignment for Reinforcement Learning With Sparse Binary Rewards
by: Minah Seo, et al.
Published: (2019-01-01) -
Combining Subgoal Graphs with Reinforcement Learning to Build a Rational Pathfinder
by: Junjie Zeng, et al.
Published: (2019-01-01) -
Learning Reward Function with Matching Network for Mapless Navigation
by: Qichen Zhang, et al.
Published: (2020-06-01) -
UCAV Air Combat Maneuver Decisions Based on a Proximal Policy Optimization Algorithm with Situation Reward Shaping
by: Kaibiao Yang, et al.
Published: (2022-08-01)