Policy Improvement for POMDPs Using Normalized Importance Sampling
We present a new method for estimating the expected return of a POMDP from experience. The estimator does not assume any knowle ge of the POMDP and allows the experience to be gathered with an arbitrary set of policies. The return is estimated for any new policy of the POMDP. We motivate the estimat...
Main Author: | Shelton, Christian R. |
---|---|
Language: | en_US |
Published: |
2004
|
Online Access: | http://hdl.handle.net/1721.1/7218 |
Similar Items
-
Sampling-based algorithms for continuous-time POMDPs
by: Chaudhari, Pratik Anil, et al.
Published: (2013) -
Stick-breaking policy learning in Dec-POMDPs
by: Amato, Christopher, et al.
Published: (2016) -
Policy Evaluation in Decentralized POMDPs With Belief Sharing
by: Mert Kayaalp, et al.
Published: (2023-01-01) -
An online algorithm for constrained POMDPs
by: Undurti, Aditya, et al.
Published: (2011) -
Point-Based Policy Transformation: Adapting Policy to Changing POMDP Models
by: Kurniawati, Hanna, et al.
Published: (2019)