Scheduled Curiosity-Deep Dyna-Q: Efficient Exploration for Dialog Policy Learning
Training task-oriented dialog agents based on reinforcement learning is time-consuming and requires a large number of interactions with real users. How to grasp dialog policy within limited dialog experiences remains an obstacle that makes the agent training process less efficient. In addition, most...
Main Authors: | Xuecheng Niu, Akinori Ito, Takashi Nose |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2024-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/10468605/ |
Similar Items
-
A Replaceable Curiosity-Driven Candidate Agent Exploration Approach for Task-Oriented Dialog Policy Learning
by: Xuecheng Niu, et al.
Published: (2024-01-01) -
Deep Dyna-Q for Rapid Learning and Improved Formation Achievement in Cooperative Transportation
by: Almira Budiyanto, et al.
Published: (2023-07-01) -
Bayesian Q learning method with Dyna architecture and prioritized sweeping
by: Jun YU, et al.
Published: (2013-11-01) -
Improved Dyna-Q: A Reinforcement Learning Method Focused via Heuristic Graph for AGV Path Planning in Dynamic Environments
by: Yiyang Liu, et al.
Published: (2022-11-01) -
Uncertainty Aware T2SS Based Dyna-Q-Learning Framework for Task Scheduling in Grid Computing
by: Bhargavi K., et al.
Published: (2022-09-01)