Optimized-Weighted-Speedy Q-Learning Algorithm for Multi-UGV in Static Environment Path Planning under Anti-Collision Cooperation Mechanism

With the accelerated development of smart cities, the concept of a “smart industrial park” in which unmanned ground vehicles (UGVs) have wide application has entered the industrial field of vision. When faced with multiple tasks and heterogeneous tasks, the task execution efficiency of a single UGV...

Full description

Bibliographic Details
Main Authors: Yuanying Cao, Xi Fang
Format: Article
Language:English
Published: MDPI AG 2023-05-01
Series:Mathematics
Subjects:
Online Access:https://www.mdpi.com/2227-7390/11/11/2476
_version_ 1797597116079538176
author Yuanying Cao
Xi Fang
author_facet Yuanying Cao
Xi Fang
author_sort Yuanying Cao
collection DOAJ
description With the accelerated development of smart cities, the concept of a “smart industrial park” in which unmanned ground vehicles (UGVs) have wide application has entered the industrial field of vision. When faced with multiple tasks and heterogeneous tasks, the task execution efficiency of a single UGV is inefficient, thus the task planning research under multi-UGV cooperation has become more urgent. In this paper, under the anti-collision cooperation mechanism for multi-UGV path planning, an improved algorithm with optimized-weighted-speedy Q-learning (OWS Q-learning) is proposed. The slow convergence speed of the Q-learning algorithm is overcome to a certain extent by changing the update mode of the Q function. By improving the selection mode of learning rate and the selection strategy of action, the relationship between exploration and utilization is balanced, and the learning efficiency of multi-agent in complex environments is improved. The simulation experiments in static environment show that the designed anti-collision coordination mechanism effectively solves the coordination problem of multiple UGVs in the same scenario. In the same experimental scenario, compared with the Q-learning algorithm and other reinforcement learning algorithms, only the OWS Q-learning algorithm achieves the convergence effect, and the OWS Q-learning algorithm has the shortest collision-free path for UGVS and the least time to complete the planning. Compared with the Q-learning algorithm, the calculation time of the OWS Q-learning algorithm in the three experimental scenarios is improved by 53.93%, 67.21%, and 53.53%, respectively. This effectively improves the intelligent development of UGV in smart parks.
first_indexed 2024-03-11T03:02:15Z
format Article
id doaj.art-03b5562e364646eba3f4d3e302bb5dff
institution Directory Open Access Journal
issn 2227-7390
language English
last_indexed 2024-03-11T03:02:15Z
publishDate 2023-05-01
publisher MDPI AG
record_format Article
series Mathematics
spelling doaj.art-03b5562e364646eba3f4d3e302bb5dff2023-11-18T08:12:33ZengMDPI AGMathematics2227-73902023-05-011111247610.3390/math11112476Optimized-Weighted-Speedy Q-Learning Algorithm for Multi-UGV in Static Environment Path Planning under Anti-Collision Cooperation MechanismYuanying Cao0Xi Fang1School of Science, Wuhan University of Technology, Wuhan 430070, ChinaSchool of Science, Wuhan University of Technology, Wuhan 430070, ChinaWith the accelerated development of smart cities, the concept of a “smart industrial park” in which unmanned ground vehicles (UGVs) have wide application has entered the industrial field of vision. When faced with multiple tasks and heterogeneous tasks, the task execution efficiency of a single UGV is inefficient, thus the task planning research under multi-UGV cooperation has become more urgent. In this paper, under the anti-collision cooperation mechanism for multi-UGV path planning, an improved algorithm with optimized-weighted-speedy Q-learning (OWS Q-learning) is proposed. The slow convergence speed of the Q-learning algorithm is overcome to a certain extent by changing the update mode of the Q function. By improving the selection mode of learning rate and the selection strategy of action, the relationship between exploration and utilization is balanced, and the learning efficiency of multi-agent in complex environments is improved. The simulation experiments in static environment show that the designed anti-collision coordination mechanism effectively solves the coordination problem of multiple UGVs in the same scenario. In the same experimental scenario, compared with the Q-learning algorithm and other reinforcement learning algorithms, only the OWS Q-learning algorithm achieves the convergence effect, and the OWS Q-learning algorithm has the shortest collision-free path for UGVS and the least time to complete the planning. Compared with the Q-learning algorithm, the calculation time of the OWS Q-learning algorithm in the three experimental scenarios is improved by 53.93%, 67.21%, and 53.53%, respectively. This effectively improves the intelligent development of UGV in smart parks.https://www.mdpi.com/2227-7390/11/11/2476optimized-weighted-speedy Q-learning algorithmpath planninganti-collision cooperation mechanismreinforcement learningunmanned ground vehicle (UGV)
spellingShingle Yuanying Cao
Xi Fang
Optimized-Weighted-Speedy Q-Learning Algorithm for Multi-UGV in Static Environment Path Planning under Anti-Collision Cooperation Mechanism
Mathematics
optimized-weighted-speedy Q-learning algorithm
path planning
anti-collision cooperation mechanism
reinforcement learning
unmanned ground vehicle (UGV)
title Optimized-Weighted-Speedy Q-Learning Algorithm for Multi-UGV in Static Environment Path Planning under Anti-Collision Cooperation Mechanism
title_full Optimized-Weighted-Speedy Q-Learning Algorithm for Multi-UGV in Static Environment Path Planning under Anti-Collision Cooperation Mechanism
title_fullStr Optimized-Weighted-Speedy Q-Learning Algorithm for Multi-UGV in Static Environment Path Planning under Anti-Collision Cooperation Mechanism
title_full_unstemmed Optimized-Weighted-Speedy Q-Learning Algorithm for Multi-UGV in Static Environment Path Planning under Anti-Collision Cooperation Mechanism
title_short Optimized-Weighted-Speedy Q-Learning Algorithm for Multi-UGV in Static Environment Path Planning under Anti-Collision Cooperation Mechanism
title_sort optimized weighted speedy q learning algorithm for multi ugv in static environment path planning under anti collision cooperation mechanism
topic optimized-weighted-speedy Q-learning algorithm
path planning
anti-collision cooperation mechanism
reinforcement learning
unmanned ground vehicle (UGV)
url https://www.mdpi.com/2227-7390/11/11/2476
work_keys_str_mv AT yuanyingcao optimizedweightedspeedyqlearningalgorithmformultiugvinstaticenvironmentpathplanningunderanticollisioncooperationmechanism
AT xifang optimizedweightedspeedyqlearningalgorithmformultiugvinstaticenvironmentpathplanningunderanticollisioncooperationmechanism