UAV Path Planning Optimization Strategy: Considerations of Urban Morphology, Microclimate, and Energy Efficiency Using Q-Learning Algorithm
The use of unmanned aerial vehicles (UAVS) has been suggested as a potential communications alternative due to their fast implantation, which makes this resource an ideal solution to provide support in scenarios such as natural disasters or intentional attacks that may cause partial or complete disr...
Main Authors: | , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2023-02-01
|
Series: | Drones |
Subjects: | |
Online Access: | https://www.mdpi.com/2504-446X/7/2/123 |
_version_ | 1797621426220433408 |
---|---|
author | Anderson Souto Rodrigo Alfaia Evelin Cardoso Jasmine Araújo Carlos Francês |
author_facet | Anderson Souto Rodrigo Alfaia Evelin Cardoso Jasmine Araújo Carlos Francês |
author_sort | Anderson Souto |
collection | DOAJ |
description | The use of unmanned aerial vehicles (UAVS) has been suggested as a potential communications alternative due to their fast implantation, which makes this resource an ideal solution to provide support in scenarios such as natural disasters or intentional attacks that may cause partial or complete disruption of telecommunications services. However, one limitation of this solution is energy autonomy, which affects mission life. With this in mind, our group has developed a new method based on reinforcement learning that aims to reduce the power consumption of UAV missions in disaster scenarios to circumvent the negative effects of wind variations, thus optimizing the timing of the aerial mesh in locations affected by the disruption of fiber-optic-based telecommunications. The method considers the K-means to stagger the position of the resource stations—from which the UAVS launched—within the topology of Stockholm, Sweden. For the UAVS’ locomotion, the Q-learning approach was used to investigate possible actions that the UAVS could take due to urban obstacles randomly distributed in the scenario and due to wind speed. The latter is related to the way the UAVS are arranged during the mission. The numerical results of the simulations have shown that the solution based on reinforcement learning was able to reduce the power consumption by 15.93% compared to the naive solution, which can lead to an increase in the life of UAV missions. |
first_indexed | 2024-03-11T08:55:46Z |
format | Article |
id | doaj.art-3d0be38f6ac44e569332d98dc7a782b0 |
institution | Directory Open Access Journal |
issn | 2504-446X |
language | English |
last_indexed | 2024-03-11T08:55:46Z |
publishDate | 2023-02-01 |
publisher | MDPI AG |
record_format | Article |
series | Drones |
spelling | doaj.art-3d0be38f6ac44e569332d98dc7a782b02023-11-16T20:06:56ZengMDPI AGDrones2504-446X2023-02-017212310.3390/drones7020123UAV Path Planning Optimization Strategy: Considerations of Urban Morphology, Microclimate, and Energy Efficiency Using Q-Learning AlgorithmAnderson Souto0Rodrigo Alfaia1Evelin Cardoso2Jasmine Araújo3Carlos Francês4Postgraduate Program in Electrical Engineering, Federal University of Pará (UFPA), Belém 66075110, BrazilPostgraduate Program in Electrical Engineering, Federal University of Pará (UFPA), Belém 66075110, BrazilPostgraduate Program in Electrical Engineering, Federal University of Pará (UFPA), Belém 66075110, BrazilPostgraduate Program in Electrical Engineering, Federal University of Pará (UFPA), Belém 66075110, BrazilPostgraduate Program in Electrical Engineering, Federal University of Pará (UFPA), Belém 66075110, BrazilThe use of unmanned aerial vehicles (UAVS) has been suggested as a potential communications alternative due to their fast implantation, which makes this resource an ideal solution to provide support in scenarios such as natural disasters or intentional attacks that may cause partial or complete disruption of telecommunications services. However, one limitation of this solution is energy autonomy, which affects mission life. With this in mind, our group has developed a new method based on reinforcement learning that aims to reduce the power consumption of UAV missions in disaster scenarios to circumvent the negative effects of wind variations, thus optimizing the timing of the aerial mesh in locations affected by the disruption of fiber-optic-based telecommunications. The method considers the K-means to stagger the position of the resource stations—from which the UAVS launched—within the topology of Stockholm, Sweden. For the UAVS’ locomotion, the Q-learning approach was used to investigate possible actions that the UAVS could take due to urban obstacles randomly distributed in the scenario and due to wind speed. The latter is related to the way the UAVS are arranged during the mission. The numerical results of the simulations have shown that the solution based on reinforcement learning was able to reduce the power consumption by 15.93% compared to the naive solution, which can lead to an increase in the life of UAV missions.https://www.mdpi.com/2504-446X/7/2/123UAVsoptimizationmachine learningQ-learningwind speedenergy consumption |
spellingShingle | Anderson Souto Rodrigo Alfaia Evelin Cardoso Jasmine Araújo Carlos Francês UAV Path Planning Optimization Strategy: Considerations of Urban Morphology, Microclimate, and Energy Efficiency Using Q-Learning Algorithm Drones UAVs optimization machine learning Q-learning wind speed energy consumption |
title | UAV Path Planning Optimization Strategy: Considerations of Urban Morphology, Microclimate, and Energy Efficiency Using Q-Learning Algorithm |
title_full | UAV Path Planning Optimization Strategy: Considerations of Urban Morphology, Microclimate, and Energy Efficiency Using Q-Learning Algorithm |
title_fullStr | UAV Path Planning Optimization Strategy: Considerations of Urban Morphology, Microclimate, and Energy Efficiency Using Q-Learning Algorithm |
title_full_unstemmed | UAV Path Planning Optimization Strategy: Considerations of Urban Morphology, Microclimate, and Energy Efficiency Using Q-Learning Algorithm |
title_short | UAV Path Planning Optimization Strategy: Considerations of Urban Morphology, Microclimate, and Energy Efficiency Using Q-Learning Algorithm |
title_sort | uav path planning optimization strategy considerations of urban morphology microclimate and energy efficiency using q learning algorithm |
topic | UAVs optimization machine learning Q-learning wind speed energy consumption |
url | https://www.mdpi.com/2504-446X/7/2/123 |
work_keys_str_mv | AT andersonsouto uavpathplanningoptimizationstrategyconsiderationsofurbanmorphologymicroclimateandenergyefficiencyusingqlearningalgorithm AT rodrigoalfaia uavpathplanningoptimizationstrategyconsiderationsofurbanmorphologymicroclimateandenergyefficiencyusingqlearningalgorithm AT evelincardoso uavpathplanningoptimizationstrategyconsiderationsofurbanmorphologymicroclimateandenergyefficiencyusingqlearningalgorithm AT jasminearaujo uavpathplanningoptimizationstrategyconsiderationsofurbanmorphologymicroclimateandenergyefficiencyusingqlearningalgorithm AT carlosfrances uavpathplanningoptimizationstrategyconsiderationsofurbanmorphologymicroclimateandenergyefficiencyusingqlearningalgorithm |