Predicting optimal value functions by interpolating reward functions in scalarized multi-objective reinforcement learning
© 2020 IEEE. A common approach for defining a reward function for multi-objective reinforcement learning (MORL) problems is the weighted sum of the multiple objectives. The weights are then treated as design parameters dependent on the expertise (and preference) of the person performing the learning...
Main Authors: | Kusari, Arpan, How, Jonathan P. |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2021
|
Online Access: | https://hdl.handle.net/1721.1/136715 |
Similar Items
-
Predicting optimal value functions by interpolating reward functions in scalarized multi-objective reinforcement learning
by: Kusari, Arpan, et al.
Published: (2021) -
Deep Reinforcement Learning With Optimized Reward Functions for Robotic Trajectory Planning
by: Jiexin Xie, et al.
Published: (2019-01-01) -
Actively learning costly reward functions for reinforcement learning
by: André Eberhard, et al.
Published: (2024-01-01) -
Inverse reinforcement learning with locally consistent reward functions
by: Nguyen, Quoc Phong, et al.
Published: (2018) -
Interpolation of functions /
by: 317613 Szabados, J., et al.
Published: (1990)