Reinforcement learning with limited reinforcement: Using Bayes risk for active learning in POMDPs
Acting in domains where an agent must plan several steps ahead to achieve a goal can be a challenging task, especially if the agentʼs sensors provide only noisy or partial information. In this setting, Partially Observable Markov Decision Processes (POMDPs) provide a planning framework that optimall...
Main Authors: | Pineau, Joelle, Doshi-Velez, Finale P, Roy, Nicholas |
---|---|
Other Authors: | Massachusetts Institute of Technology. Department of Aeronautics and Astronautics |
Format: | Article |
Language: | en_US |
Published: |
Elsevier
2017
|
Online Access: | http://hdl.handle.net/1721.1/108303 https://orcid.org/0000-0002-8293-0492 |
Similar Items
-
Deep variational reinforcement learning for POMDPs
by: Igl, M, et al.
Published: (2018) -
Bayesian Nonparametric Methods for Partially-Observable Reinforcement Learning
by: Doshi-Velez, Finale P., et al.
Published: (2015) -
Nonparametric Bayesian Policy Priors for Reinforcement Learning
by: Doshi-Velez, Finale P., et al.
Published: (2011) -
Multi-Agent Active Perception Based on Reinforcement Learning and POMDP
by: Tarik Selimovic, et al.
Published: (2024-01-01) -
Bayesian nonparametric approaches for reinforcement learning in partially observable domains
by: Doshi-Velez, Finale
Published: (2012)