A Sample Aggregation Approach to Experiences Replay of Dyna-Q Learning
In a complex environment, the learning efficiency of reinforcement learning methods always decreases due to large-scale or continuous spaces problems, which can cause the well-known curse of dimensionality. To deal with this problem and enhance learning efficiency, this paper introduces an aggregati...
Main Authors: | Haobin Shi, Shike Yang, Kao-Shing Hwang, Jialin Chen, Mengkai Hu, Hengsheng Zhang |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2018-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/8383982/ |
Similar Items
-
Deep Dyna-Q for Rapid Learning and Improved Formation Achievement in Cooperative Transportation
by: Almira Budiyanto, et al.
Published: (2023-07-01) -
A Dyna-Q-Based Solution for UAV Networks Against Smart Jamming Attacks
by: Zhiwei Li, et al.
Published: (2019-05-01) -
Scheduled Curiosity-Deep Dyna-Q: Efficient Exploration for Dialog Policy Learning
by: Xuecheng Niu, et al.
Published: (2024-01-01) -
Uncertainty Aware T2SS Based Dyna-Q-Learning Framework for Task Scheduling in Grid Computing
by: Bhargavi K., et al.
Published: (2022-09-01) -
STAT: a fast, scalable, MinHash-based k-mer tool to assess Sequence Read Archive next-generation sequence submissions
by: Kenneth S. Katz, et al.
Published: (2021-09-01)