Fourier policy gradients
We propose a new way of deriving policy gradient updates for reinforcement learning. Our technique, based on Fourier analysis, recasts integrals that arise with expected policy gradients as convolutions and turns them into multiplications. The obtained analytical solutions allow us to capture the lo...
主要な著者: | Fellows, M, Ciosek, K, Whiteson, S |
---|---|
フォーマット: | Conference item |
出版事項: |
Journal of Machine Learning Research
2018
|
類似資料
-
Expected policy gradients
著者:: Ciosek, K, 等
出版事項: (2018) -
Expected policy gradients for reinforcement learning
著者:: Ciosek, K, 等
出版事項: (2020) -
OFFER: Off-environment reinforcement learning
著者:: Ciosek, K, 等
出版事項: (2017) -
Counterfactual multi−agent policy gradients
著者:: Foerster, J, 等
出版事項: (2018) -
Fast efficient hyperparameter tuning for policy gradient methods
著者:: Paul, S, 等
出版事項: (2019)