A Reinforcement Learning Method Based on an Improved Sampling Mechanism for Unmanned Aerial Vehicle Penetration

The penetration of unmanned aerial vehicles (UAVs) is an important aspect of UAV games. In recent years, UAV penetration has generally been solved using artificial intelligence methods such as reinforcement learning. However, the high sample demand of the reinforcement learning method poses a signif...

Full description

Bibliographic Details
Main Authors: Yue Wang, Kexv Li, Xing Zhuang, Xinyu Liu, Hanyu Li
Format: Article
Language:English
Published: MDPI AG 2023-07-01
Series:Aerospace
Subjects:
Online Access:https://www.mdpi.com/2226-4310/10/7/642
_version_ 1797590743983849472
author Yue Wang
Kexv Li
Xing Zhuang
Xinyu Liu
Hanyu Li
author_facet Yue Wang
Kexv Li
Xing Zhuang
Xinyu Liu
Hanyu Li
author_sort Yue Wang
collection DOAJ
description The penetration of unmanned aerial vehicles (UAVs) is an important aspect of UAV games. In recent years, UAV penetration has generally been solved using artificial intelligence methods such as reinforcement learning. However, the high sample demand of the reinforcement learning method poses a significant challenge specifically in the context of UAV games. To improve the sample utilization in UAV penetration, this paper innovatively proposes an improved sampling mechanism called task completion division (TCD) and combines this method with the soft actor critic (SAC) algorithm to form the TCD-SAC algorithm. To compare the performance of the TCD-SAC algorithm with other related baseline algorithms, this study builds a dynamic environment, a UAV game, and conducts training and testing experiments in this environment. The results show that among all the algorithms, the TCD-SAC algorithm has the highest sample utilization rate and the best actual penetration results, and the algorithm has a good adaptability and robustness in dynamic environments.
first_indexed 2024-03-11T01:24:50Z
format Article
id doaj.art-c40ce32ee61744eeb86cb440dd991e15
institution Directory Open Access Journal
issn 2226-4310
language English
last_indexed 2024-03-11T01:24:50Z
publishDate 2023-07-01
publisher MDPI AG
record_format Article
series Aerospace
spelling doaj.art-c40ce32ee61744eeb86cb440dd991e152023-11-18T17:51:12ZengMDPI AGAerospace2226-43102023-07-0110764210.3390/aerospace10070642A Reinforcement Learning Method Based on an Improved Sampling Mechanism for Unmanned Aerial Vehicle PenetrationYue Wang0Kexv Li1Xing Zhuang2Xinyu Liu3Hanyu Li4School of Mechatronical Engineering, Beijing Institute of Technology, Beijing 100081, ChinaSchool of Mechatronical Engineering, Beijing Institute of Technology, Beijing 100081, ChinaSchool of Mechatronical Engineering, Beijing Institute of Technology, Beijing 100081, ChinaSchool of Mechatronical Engineering, Beijing Institute of Technology, Beijing 100081, ChinaSchool of Mechatronical Engineering, Beijing Institute of Technology, Beijing 100081, ChinaThe penetration of unmanned aerial vehicles (UAVs) is an important aspect of UAV games. In recent years, UAV penetration has generally been solved using artificial intelligence methods such as reinforcement learning. However, the high sample demand of the reinforcement learning method poses a significant challenge specifically in the context of UAV games. To improve the sample utilization in UAV penetration, this paper innovatively proposes an improved sampling mechanism called task completion division (TCD) and combines this method with the soft actor critic (SAC) algorithm to form the TCD-SAC algorithm. To compare the performance of the TCD-SAC algorithm with other related baseline algorithms, this study builds a dynamic environment, a UAV game, and conducts training and testing experiments in this environment. The results show that among all the algorithms, the TCD-SAC algorithm has the highest sample utilization rate and the best actual penetration results, and the algorithm has a good adaptability and robustness in dynamic environments.https://www.mdpi.com/2226-4310/10/7/642UAV penetrationreinforcement learningsample utilizationtask completion division
spellingShingle Yue Wang
Kexv Li
Xing Zhuang
Xinyu Liu
Hanyu Li
A Reinforcement Learning Method Based on an Improved Sampling Mechanism for Unmanned Aerial Vehicle Penetration
Aerospace
UAV penetration
reinforcement learning
sample utilization
task completion division
title A Reinforcement Learning Method Based on an Improved Sampling Mechanism for Unmanned Aerial Vehicle Penetration
title_full A Reinforcement Learning Method Based on an Improved Sampling Mechanism for Unmanned Aerial Vehicle Penetration
title_fullStr A Reinforcement Learning Method Based on an Improved Sampling Mechanism for Unmanned Aerial Vehicle Penetration
title_full_unstemmed A Reinforcement Learning Method Based on an Improved Sampling Mechanism for Unmanned Aerial Vehicle Penetration
title_short A Reinforcement Learning Method Based on an Improved Sampling Mechanism for Unmanned Aerial Vehicle Penetration
title_sort reinforcement learning method based on an improved sampling mechanism for unmanned aerial vehicle penetration
topic UAV penetration
reinforcement learning
sample utilization
task completion division
url https://www.mdpi.com/2226-4310/10/7/642
work_keys_str_mv AT yuewang areinforcementlearningmethodbasedonanimprovedsamplingmechanismforunmannedaerialvehiclepenetration
AT kexvli areinforcementlearningmethodbasedonanimprovedsamplingmechanismforunmannedaerialvehiclepenetration
AT xingzhuang areinforcementlearningmethodbasedonanimprovedsamplingmechanismforunmannedaerialvehiclepenetration
AT xinyuliu areinforcementlearningmethodbasedonanimprovedsamplingmechanismforunmannedaerialvehiclepenetration
AT hanyuli areinforcementlearningmethodbasedonanimprovedsamplingmechanismforunmannedaerialvehiclepenetration
AT yuewang reinforcementlearningmethodbasedonanimprovedsamplingmechanismforunmannedaerialvehiclepenetration
AT kexvli reinforcementlearningmethodbasedonanimprovedsamplingmechanismforunmannedaerialvehiclepenetration
AT xingzhuang reinforcementlearningmethodbasedonanimprovedsamplingmechanismforunmannedaerialvehiclepenetration
AT xinyuliu reinforcementlearningmethodbasedonanimprovedsamplingmechanismforunmannedaerialvehiclepenetration
AT hanyuli reinforcementlearningmethodbasedonanimprovedsamplingmechanismforunmannedaerialvehiclepenetration