Obstacle avoidance USV in multi-static obstacle environments based on a deep reinforcement learning approach

Unmanned surface vehicles (USVs) are intelligent platforms for unmanned surface navigation based on artificial intelligence, motion control, environmental awareness, and other professional technologies. Obstacle avoidance is an important part of its autonomous navigation. Although the USV works in t...

Full description

Bibliographic Details
Main Authors: Dengyao Jiang, Mingzhe Yuan, Junfeng Xiong, Jinchao Xiao, Yong Duan
Format: Article
Language:English
Published: SAGE Publishing 2024-04-01
Series:Measurement + Control
Online Access:https://doi.org/10.1177/00202940231195937
_version_ 1797262316029345792
author Dengyao Jiang
Mingzhe Yuan
Junfeng Xiong
Jinchao Xiao
Yong Duan
author_facet Dengyao Jiang
Mingzhe Yuan
Junfeng Xiong
Jinchao Xiao
Yong Duan
author_sort Dengyao Jiang
collection DOAJ
description Unmanned surface vehicles (USVs) are intelligent platforms for unmanned surface navigation based on artificial intelligence, motion control, environmental awareness, and other professional technologies. Obstacle avoidance is an important part of its autonomous navigation. Although the USV works in the water environment (e.g. monitoring and tracking, search and rescue scenarios), the dynamic and complex operating environment makes the traditional methods not suitable for solving the obstacle avoidance problem of the USV. In this paper, to address the issue of poor convergence of the Twin Delayed Deep Deterministic policy gradient (TD3) algorithm of Deep Reinforcement Learning (DRL) in an unstructured environment and wave current interference, random walk policy is proposed to deposit the pre-exploration policy of the algorithm into the experience pool to accelerate the convergence of the algorithm and thus achieve USV obstacle avoidance, which can achieve collision-free navigation from any start point to a given end point in a dynamic and complex environment without offline trajectory and track point generation. We design a pre-exploration policy for the environment and a virtual simulation environment for training and testing the algorithm and give the reward function and training method. The simulation results show that our proposed algorithm is more manageable to converge than the original algorithm and can perform better in complex environments in terms of obstacle avoidance behavior, reflecting the algorithm’s feasibility and effectiveness.
first_indexed 2024-04-24T23:55:10Z
format Article
id doaj.art-80049420927642ddbc8a55d5b7c44032
institution Directory Open Access Journal
issn 0020-2940
language English
last_indexed 2024-04-24T23:55:10Z
publishDate 2024-04-01
publisher SAGE Publishing
record_format Article
series Measurement + Control
spelling doaj.art-80049420927642ddbc8a55d5b7c440322024-03-14T14:03:34ZengSAGE PublishingMeasurement + Control0020-29402024-04-015710.1177/00202940231195937Obstacle avoidance USV in multi-static obstacle environments based on a deep reinforcement learning approachDengyao Jiang0Mingzhe Yuan1Junfeng Xiong2Jinchao Xiao3Yong Duan4Guangzhou Institute of Industrial Intelligence, Guangzhou, ChinaAcademicians Experts Workstation, Guangzhou Institute of Industrial Intelligence, Guangzhou, ChinaAcademicians Experts Workstation, Guangzhou Institute of Industrial Intelligence, Guangzhou, ChinaAcademicians Experts Workstation, Guangzhou Institute of Industrial Intelligence, Guangzhou, ChinaSchool of Information Science and Engineering, Shenyang University of Technology, Shenyang, ChinaUnmanned surface vehicles (USVs) are intelligent platforms for unmanned surface navigation based on artificial intelligence, motion control, environmental awareness, and other professional technologies. Obstacle avoidance is an important part of its autonomous navigation. Although the USV works in the water environment (e.g. monitoring and tracking, search and rescue scenarios), the dynamic and complex operating environment makes the traditional methods not suitable for solving the obstacle avoidance problem of the USV. In this paper, to address the issue of poor convergence of the Twin Delayed Deep Deterministic policy gradient (TD3) algorithm of Deep Reinforcement Learning (DRL) in an unstructured environment and wave current interference, random walk policy is proposed to deposit the pre-exploration policy of the algorithm into the experience pool to accelerate the convergence of the algorithm and thus achieve USV obstacle avoidance, which can achieve collision-free navigation from any start point to a given end point in a dynamic and complex environment without offline trajectory and track point generation. We design a pre-exploration policy for the environment and a virtual simulation environment for training and testing the algorithm and give the reward function and training method. The simulation results show that our proposed algorithm is more manageable to converge than the original algorithm and can perform better in complex environments in terms of obstacle avoidance behavior, reflecting the algorithm’s feasibility and effectiveness.https://doi.org/10.1177/00202940231195937
spellingShingle Dengyao Jiang
Mingzhe Yuan
Junfeng Xiong
Jinchao Xiao
Yong Duan
Obstacle avoidance USV in multi-static obstacle environments based on a deep reinforcement learning approach
Measurement + Control
title Obstacle avoidance USV in multi-static obstacle environments based on a deep reinforcement learning approach
title_full Obstacle avoidance USV in multi-static obstacle environments based on a deep reinforcement learning approach
title_fullStr Obstacle avoidance USV in multi-static obstacle environments based on a deep reinforcement learning approach
title_full_unstemmed Obstacle avoidance USV in multi-static obstacle environments based on a deep reinforcement learning approach
title_short Obstacle avoidance USV in multi-static obstacle environments based on a deep reinforcement learning approach
title_sort obstacle avoidance usv in multi static obstacle environments based on a deep reinforcement learning approach
url https://doi.org/10.1177/00202940231195937
work_keys_str_mv AT dengyaojiang obstacleavoidanceusvinmultistaticobstacleenvironmentsbasedonadeepreinforcementlearningapproach
AT mingzheyuan obstacleavoidanceusvinmultistaticobstacleenvironmentsbasedonadeepreinforcementlearningapproach
AT junfengxiong obstacleavoidanceusvinmultistaticobstacleenvironmentsbasedonadeepreinforcementlearningapproach
AT jinchaoxiao obstacleavoidanceusvinmultistaticobstacleenvironmentsbasedonadeepreinforcementlearningapproach
AT yongduan obstacleavoidanceusvinmultistaticobstacleenvironmentsbasedonadeepreinforcementlearningapproach