A new noise network and gradient parallelisation‐based asynchronous advantage actor‐critic algorithm
Abstract Asynchronous advantage actor‐critic (A3C) algorithm is a commonly used policy optimization algorithm in reinforcement learning, in which asynchronous is parallel interactive sampling and training, and advantage is a sampling multi‐step reward estimation method for computing weights. In orde...
Main Authors: | , , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Wiley
2022-09-01
|
Series: | IET Cyber-systems and Robotics |
Subjects: | |
Online Access: | https://doi.org/10.1049/csy2.12059 |
_version_ | 1811254379487952896 |
---|---|
author | Zhengshun Fei Yanping Wang Jinglong Wang Kangling Liu Bingqiang Huang Ping Tan |
author_facet | Zhengshun Fei Yanping Wang Jinglong Wang Kangling Liu Bingqiang Huang Ping Tan |
author_sort | Zhengshun Fei |
collection | DOAJ |
description | Abstract Asynchronous advantage actor‐critic (A3C) algorithm is a commonly used policy optimization algorithm in reinforcement learning, in which asynchronous is parallel interactive sampling and training, and advantage is a sampling multi‐step reward estimation method for computing weights. In order to address the problem of low efficiency and insufficient convergence caused by the traditional heuristic exploration of A3C algorithm in reinforcement learning, an improved A3C algorithm is proposed in this paper. In this algorithm, a noise network function, which updates the noise tensor in an explicit way is constructed to train the agent. Generalised advantage estimation (GAE) is also adopted to describe the dominance function. Finally, a new mean gradient parallelisation method is designed to update the parameters in both the primary and secondary networks by summing and averaging the gradients passed from all the sub‐processes to the main process. Simulation experiments were conducted in a gym environment using the PyTorch Agent Net (PTAN) advanced reinforcement learning library, and the results show that the method enables the agent to complete the learning training faster and its convergence during the training process is better. The improved A3C algorithm has a better performance than the original algorithm, which can provide new ideas for subsequent research on reinforcement learning algorithms. |
first_indexed | 2024-04-12T17:06:23Z |
format | Article |
id | doaj.art-aa080922baa346f999cf7aa32107eee0 |
institution | Directory Open Access Journal |
issn | 2631-6315 |
language | English |
last_indexed | 2024-04-12T17:06:23Z |
publishDate | 2022-09-01 |
publisher | Wiley |
record_format | Article |
series | IET Cyber-systems and Robotics |
spelling | doaj.art-aa080922baa346f999cf7aa32107eee02022-12-22T03:23:56ZengWileyIET Cyber-systems and Robotics2631-63152022-09-014317518810.1049/csy2.12059A new noise network and gradient parallelisation‐based asynchronous advantage actor‐critic algorithmZhengshun Fei0Yanping Wang1Jinglong Wang2Kangling Liu3Bingqiang Huang4Ping Tan5Provincial Key Institute of Robotics School of Automation and Electrical Engineering Zhejiang University of Science and Technology Hangzhou ChinaProvincial Key Institute of Robotics School of Automation and Electrical Engineering Zhejiang University of Science and Technology Hangzhou ChinaProvincial Key Institute of Robotics School of Automation and Electrical Engineering Zhejiang University of Science and Technology Hangzhou ChinaState Key Lab of Industrial Control Technology College of Control Science and Engineering Zhejiang University Hangzhou ChinaProvincial Key Institute of Robotics School of Automation and Electrical Engineering Zhejiang University of Science and Technology Hangzhou ChinaProvincial Key Institute of Robotics School of Automation and Electrical Engineering Zhejiang University of Science and Technology Hangzhou ChinaAbstract Asynchronous advantage actor‐critic (A3C) algorithm is a commonly used policy optimization algorithm in reinforcement learning, in which asynchronous is parallel interactive sampling and training, and advantage is a sampling multi‐step reward estimation method for computing weights. In order to address the problem of low efficiency and insufficient convergence caused by the traditional heuristic exploration of A3C algorithm in reinforcement learning, an improved A3C algorithm is proposed in this paper. In this algorithm, a noise network function, which updates the noise tensor in an explicit way is constructed to train the agent. Generalised advantage estimation (GAE) is also adopted to describe the dominance function. Finally, a new mean gradient parallelisation method is designed to update the parameters in both the primary and secondary networks by summing and averaging the gradients passed from all the sub‐processes to the main process. Simulation experiments were conducted in a gym environment using the PyTorch Agent Net (PTAN) advanced reinforcement learning library, and the results show that the method enables the agent to complete the learning training faster and its convergence during the training process is better. The improved A3C algorithm has a better performance than the original algorithm, which can provide new ideas for subsequent research on reinforcement learning algorithms.https://doi.org/10.1049/csy2.12059asynchronous advantage actor‐critic (A3C)generalised advantage estimation (GAE)parallelisationreinforcement learning |
spellingShingle | Zhengshun Fei Yanping Wang Jinglong Wang Kangling Liu Bingqiang Huang Ping Tan A new noise network and gradient parallelisation‐based asynchronous advantage actor‐critic algorithm IET Cyber-systems and Robotics asynchronous advantage actor‐critic (A3C) generalised advantage estimation (GAE) parallelisation reinforcement learning |
title | A new noise network and gradient parallelisation‐based asynchronous advantage actor‐critic algorithm |
title_full | A new noise network and gradient parallelisation‐based asynchronous advantage actor‐critic algorithm |
title_fullStr | A new noise network and gradient parallelisation‐based asynchronous advantage actor‐critic algorithm |
title_full_unstemmed | A new noise network and gradient parallelisation‐based asynchronous advantage actor‐critic algorithm |
title_short | A new noise network and gradient parallelisation‐based asynchronous advantage actor‐critic algorithm |
title_sort | new noise network and gradient parallelisation based asynchronous advantage actor critic algorithm |
topic | asynchronous advantage actor‐critic (A3C) generalised advantage estimation (GAE) parallelisation reinforcement learning |
url | https://doi.org/10.1049/csy2.12059 |
work_keys_str_mv | AT zhengshunfei anewnoisenetworkandgradientparallelisationbasedasynchronousadvantageactorcriticalgorithm AT yanpingwang anewnoisenetworkandgradientparallelisationbasedasynchronousadvantageactorcriticalgorithm AT jinglongwang anewnoisenetworkandgradientparallelisationbasedasynchronousadvantageactorcriticalgorithm AT kanglingliu anewnoisenetworkandgradientparallelisationbasedasynchronousadvantageactorcriticalgorithm AT bingqianghuang anewnoisenetworkandgradientparallelisationbasedasynchronousadvantageactorcriticalgorithm AT pingtan anewnoisenetworkandgradientparallelisationbasedasynchronousadvantageactorcriticalgorithm AT zhengshunfei newnoisenetworkandgradientparallelisationbasedasynchronousadvantageactorcriticalgorithm AT yanpingwang newnoisenetworkandgradientparallelisationbasedasynchronousadvantageactorcriticalgorithm AT jinglongwang newnoisenetworkandgradientparallelisationbasedasynchronousadvantageactorcriticalgorithm AT kanglingliu newnoisenetworkandgradientparallelisationbasedasynchronousadvantageactorcriticalgorithm AT bingqianghuang newnoisenetworkandgradientparallelisationbasedasynchronousadvantageactorcriticalgorithm AT pingtan newnoisenetworkandgradientparallelisationbasedasynchronousadvantageactorcriticalgorithm |