Computationally Efficient Reinforcement Learning under Partial Observability
A key challenge in reinforcement learning is the inability of the agent to fully observe the latent state of the system. Partially observable Markov decision processes (POMDPs) are a generalization of Markov decision processes (MDPs) that model this challenge. Unfortunately, planning and learning ne...
Main Author: | Rohatgi, Dhruv |
---|---|
Other Authors: | Moitra, Ankur |
Format: | Thesis |
Published: |
Massachusetts Institute of Technology
2023
|
Online Access: | https://hdl.handle.net/1721.1/150191 |
Similar Items
-
Planning and Learning in Partially Observable Systems via Filter Stability
by: Golowich, Noah, et al.
Published: (2023) -
Exploring and Learning in Sparse Linear MDPs without Computationally Intractable Oracles
by: Golowich, Noah, et al.
Published: (2024) -
Deep decentralized multi-task multi-agent reinforcement learning under partial observability
by: How, Jonathan
Published: (2021) -
Bayesian Nonparametric Methods for Partially-Observable Reinforcement Learning
by: Doshi-Velez, Finale P., et al.
Published: (2015) -
Bayesian nonparametric approaches for reinforcement learning in partially observable domains
by: Doshi-Velez, Finale
Published: (2012)