PPO-Exp: Keeping Fixed-Wing UAV Formation with Deep Reinforcement Learning
Flocking for fixed-Wing Unmanned Aerial Vehicles (UAVs) is an extremely complex challenge due to fixed-wing UAV’s control problem and the system’s coordinate difficulty. Recently, flocking approaches based on reinforcement learning have attracted attention. However, current methods also require that...
Main Authors: | , , , , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2022-12-01
|
Series: | Drones |
Subjects: | |
Online Access: | https://www.mdpi.com/2504-446X/7/1/28 |
_version_ | 1827626690764865536 |
---|---|
author | Dan Xu Yunxiao Guo Zhongyi Yu Zhenfeng Wang Rongze Lan Runhao Zhao Xinjia Xie Han Long |
author_facet | Dan Xu Yunxiao Guo Zhongyi Yu Zhenfeng Wang Rongze Lan Runhao Zhao Xinjia Xie Han Long |
author_sort | Dan Xu |
collection | DOAJ |
description | Flocking for fixed-Wing Unmanned Aerial Vehicles (UAVs) is an extremely complex challenge due to fixed-wing UAV’s control problem and the system’s coordinate difficulty. Recently, flocking approaches based on reinforcement learning have attracted attention. However, current methods also require that each UAV makes the decision decentralized, which increases the cost and computation of the whole UAV system. This paper researches a low-cost UAV formation system consisting of one leader (equipped with the intelligence chip) with five followers (without the intelligence chip), and proposes a centralized collision-free formation-keeping method. The communication in the whole process is considered and the protocol is designed by minimizing the communication cost. In addition, an analysis of the Proximal Policy Optimization (PPO) algorithm is provided; the paper derives the estimation error bound, and reveals the relationship between the bound and exploration. To encourage the agent to balance their exploration and estimation error bound, a version of PPO named PPO-Exploration (PPO-Exp) is proposed. It can adjust the clip constraint parameter and make the exploration mechanism more flexible. The results of the experiments show that PPO-Exp performs better than the current algorithms in these tasks. |
first_indexed | 2024-03-09T13:00:09Z |
format | Article |
id | doaj.art-c6ab4f1ac76d493ba7ce24f9710da27c |
institution | Directory Open Access Journal |
issn | 2504-446X |
language | English |
last_indexed | 2024-03-09T13:00:09Z |
publishDate | 2022-12-01 |
publisher | MDPI AG |
record_format | Article |
series | Drones |
spelling | doaj.art-c6ab4f1ac76d493ba7ce24f9710da27c2023-11-30T21:55:17ZengMDPI AGDrones2504-446X2022-12-01712810.3390/drones7010028PPO-Exp: Keeping Fixed-Wing UAV Formation with Deep Reinforcement LearningDan Xu0Yunxiao Guo1Zhongyi Yu2Zhenfeng Wang3Rongze Lan4Runhao Zhao5Xinjia Xie6Han Long7College of System Engineering, National University of Defense Technology, Changsha 410073, ChinaCollege of Sciences, National University of Defense Technology, Changsha 410073, ChinaCollege of Advanced Interdisciplinary Studies, National University of Defense Technology, Changsha 410073, ChinaCollege of Sciences, National University of Defense Technology, Changsha 410073, ChinaCollege of Sciences, National University of Defense Technology, Changsha 410073, ChinaCollege of System Engineering, National University of Defense Technology, Changsha 410073, ChinaCollege of Computer Science, National University of Defense Technology, Changsha 410073, ChinaCollege of Sciences, National University of Defense Technology, Changsha 410073, ChinaFlocking for fixed-Wing Unmanned Aerial Vehicles (UAVs) is an extremely complex challenge due to fixed-wing UAV’s control problem and the system’s coordinate difficulty. Recently, flocking approaches based on reinforcement learning have attracted attention. However, current methods also require that each UAV makes the decision decentralized, which increases the cost and computation of the whole UAV system. This paper researches a low-cost UAV formation system consisting of one leader (equipped with the intelligence chip) with five followers (without the intelligence chip), and proposes a centralized collision-free formation-keeping method. The communication in the whole process is considered and the protocol is designed by minimizing the communication cost. In addition, an analysis of the Proximal Policy Optimization (PPO) algorithm is provided; the paper derives the estimation error bound, and reveals the relationship between the bound and exploration. To encourage the agent to balance their exploration and estimation error bound, a version of PPO named PPO-Exploration (PPO-Exp) is proposed. It can adjust the clip constraint parameter and make the exploration mechanism more flexible. The results of the experiments show that PPO-Exp performs better than the current algorithms in these tasks.https://www.mdpi.com/2504-446X/7/1/28fixed-wing UAVformation keepingreinforcement learning |
spellingShingle | Dan Xu Yunxiao Guo Zhongyi Yu Zhenfeng Wang Rongze Lan Runhao Zhao Xinjia Xie Han Long PPO-Exp: Keeping Fixed-Wing UAV Formation with Deep Reinforcement Learning Drones fixed-wing UAV formation keeping reinforcement learning |
title | PPO-Exp: Keeping Fixed-Wing UAV Formation with Deep Reinforcement Learning |
title_full | PPO-Exp: Keeping Fixed-Wing UAV Formation with Deep Reinforcement Learning |
title_fullStr | PPO-Exp: Keeping Fixed-Wing UAV Formation with Deep Reinforcement Learning |
title_full_unstemmed | PPO-Exp: Keeping Fixed-Wing UAV Formation with Deep Reinforcement Learning |
title_short | PPO-Exp: Keeping Fixed-Wing UAV Formation with Deep Reinforcement Learning |
title_sort | ppo exp keeping fixed wing uav formation with deep reinforcement learning |
topic | fixed-wing UAV formation keeping reinforcement learning |
url | https://www.mdpi.com/2504-446X/7/1/28 |
work_keys_str_mv | AT danxu ppoexpkeepingfixedwinguavformationwithdeepreinforcementlearning AT yunxiaoguo ppoexpkeepingfixedwinguavformationwithdeepreinforcementlearning AT zhongyiyu ppoexpkeepingfixedwinguavformationwithdeepreinforcementlearning AT zhenfengwang ppoexpkeepingfixedwinguavformationwithdeepreinforcementlearning AT rongzelan ppoexpkeepingfixedwinguavformationwithdeepreinforcementlearning AT runhaozhao ppoexpkeepingfixedwinguavformationwithdeepreinforcementlearning AT xinjiaxie ppoexpkeepingfixedwinguavformationwithdeepreinforcementlearning AT hanlong ppoexpkeepingfixedwinguavformationwithdeepreinforcementlearning |