Predicting optimal value functions by interpolating reward functions in scalarized multi-objective reinforcement learning
© 2020 IEEE. A common approach for defining a reward function for multi-objective reinforcement learning (MORL) problems is the weighted sum of the multiple objectives. The weights are then treated as design parameters dependent on the expertise (and preference) of the person performing the learning...
Main Authors: | Kusari, Arpan, How, Jonathan P. |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2021
|
Online Access: | https://hdl.handle.net/1721.1/136715 |
Similar Items
-
Predicting optimal value functions by interpolating reward functions in scalarized multi-objective reinforcement learning
by: Kusari, Arpan, et al.
Published: (2021) -
Intelligent Frequency Control Strategy Based on Reinforcement Learning of Multi-Objective Collaborative Reward Function
by: Lei Zhang, et al.
Published: (2021-09-01) -
Deep Reinforcement Learning With Optimized Reward Functions for Robotic Trajectory Planning
by: Jiexin Xie, et al.
Published: (2019-01-01) -
Inverse reinforcement learning with locally consistent reward functions
by: Nguyen, Quoc Phong, et al.
Published: (2018) -
Actively learning costly reward functions for reinforcement learning
by: André Eberhard, et al.
Published: (2024-01-01)