Deep Reinforcement Learning for End-to-End Local Motion Planning of Autonomous Aerial Robots in Unknown Outdoor Environments: Real-Time Flight Experiments

Autonomous navigation and collision avoidance missions represent a significant challenge for robotics systems as they generally operate in dynamic environments that require a high level of autonomy and flexible decision-making capabilities. This challenge becomes more applicable in micro aerial vehi...

Full description

Bibliographic Details
Main Authors: Oualid Doukhi, Deok-Jin Lee
Format: Article
Language:English
Published: MDPI AG 2021-04-01
Series:Sensors
Subjects:
Online Access:https://www.mdpi.com/1424-8220/21/7/2534
_version_ 1797538827443634176
author Oualid Doukhi
Deok-Jin Lee
author_facet Oualid Doukhi
Deok-Jin Lee
author_sort Oualid Doukhi
collection DOAJ
description Autonomous navigation and collision avoidance missions represent a significant challenge for robotics systems as they generally operate in dynamic environments that require a high level of autonomy and flexible decision-making capabilities. This challenge becomes more applicable in micro aerial vehicles (MAVs) due to their limited size and computational power. This paper presents a novel approach for enabling a micro aerial vehicle system equipped with a laser range finder to autonomously navigate among obstacles and achieve a user-specified goal location in a GPS-denied environment, without the need for mapping or path planning. The proposed system uses an actor–critic-based reinforcement learning technique to train the aerial robot in a Gazebo simulator to perform a point-goal navigation task by directly mapping the noisy MAV’s state and laser scan measurements to continuous motion control. The obtained policy can perform collision-free flight in the real world while being trained entirely on a 3D simulator. Intensive simulations and real-time experiments were conducted and compared with a nonlinear model predictive control technique to show the generalization capabilities to new unseen environments, and robustness against localization noise. The obtained results demonstrate our system’s effectiveness in flying safely and reaching the desired points by planning smooth forward linear velocity and heading rates.
first_indexed 2024-03-10T12:36:57Z
format Article
id doaj.art-ac1c6906b18449b7bacbf4486b4f4d3e
institution Directory Open Access Journal
issn 1424-8220
language English
last_indexed 2024-03-10T12:36:57Z
publishDate 2021-04-01
publisher MDPI AG
record_format Article
series Sensors
spelling doaj.art-ac1c6906b18449b7bacbf4486b4f4d3e2023-11-21T14:13:28ZengMDPI AGSensors1424-82202021-04-01217253410.3390/s21072534Deep Reinforcement Learning for End-to-End Local Motion Planning of Autonomous Aerial Robots in Unknown Outdoor Environments: Real-Time Flight ExperimentsOualid Doukhi0Deok-Jin Lee1Center for Artificial Intelligence & Autonomous Systems, Kunsan National University, 558 Daehak-ro, Naun 2(i)-dong, Gunsan 54150, Jeollabuk-do, KoreaSchool of Mechanical Design Engineering, Smart e-Mobilty Lab, Center for Artificial Intelligence & Autonomous Systems, Jeonbuk National University, 567, Baekje-daero, Deokjin-gu, Jeonju-si 54896, Jeollabuk-do, KoreaAutonomous navigation and collision avoidance missions represent a significant challenge for robotics systems as they generally operate in dynamic environments that require a high level of autonomy and flexible decision-making capabilities. This challenge becomes more applicable in micro aerial vehicles (MAVs) due to their limited size and computational power. This paper presents a novel approach for enabling a micro aerial vehicle system equipped with a laser range finder to autonomously navigate among obstacles and achieve a user-specified goal location in a GPS-denied environment, without the need for mapping or path planning. The proposed system uses an actor–critic-based reinforcement learning technique to train the aerial robot in a Gazebo simulator to perform a point-goal navigation task by directly mapping the noisy MAV’s state and laser scan measurements to continuous motion control. The obtained policy can perform collision-free flight in the real world while being trained entirely on a 3D simulator. Intensive simulations and real-time experiments were conducted and compared with a nonlinear model predictive control technique to show the generalization capabilities to new unseen environments, and robustness against localization noise. The obtained results demonstrate our system’s effectiveness in flying safely and reaching the desired points by planning smooth forward linear velocity and heading rates.https://www.mdpi.com/1424-8220/21/7/2534autonomous navigationcollision-freedeep reinforcement learningunmanned aerial vehicle
spellingShingle Oualid Doukhi
Deok-Jin Lee
Deep Reinforcement Learning for End-to-End Local Motion Planning of Autonomous Aerial Robots in Unknown Outdoor Environments: Real-Time Flight Experiments
Sensors
autonomous navigation
collision-free
deep reinforcement learning
unmanned aerial vehicle
title Deep Reinforcement Learning for End-to-End Local Motion Planning of Autonomous Aerial Robots in Unknown Outdoor Environments: Real-Time Flight Experiments
title_full Deep Reinforcement Learning for End-to-End Local Motion Planning of Autonomous Aerial Robots in Unknown Outdoor Environments: Real-Time Flight Experiments
title_fullStr Deep Reinforcement Learning for End-to-End Local Motion Planning of Autonomous Aerial Robots in Unknown Outdoor Environments: Real-Time Flight Experiments
title_full_unstemmed Deep Reinforcement Learning for End-to-End Local Motion Planning of Autonomous Aerial Robots in Unknown Outdoor Environments: Real-Time Flight Experiments
title_short Deep Reinforcement Learning for End-to-End Local Motion Planning of Autonomous Aerial Robots in Unknown Outdoor Environments: Real-Time Flight Experiments
title_sort deep reinforcement learning for end to end local motion planning of autonomous aerial robots in unknown outdoor environments real time flight experiments
topic autonomous navigation
collision-free
deep reinforcement learning
unmanned aerial vehicle
url https://www.mdpi.com/1424-8220/21/7/2534
work_keys_str_mv AT oualiddoukhi deepreinforcementlearningforendtoendlocalmotionplanningofautonomousaerialrobotsinunknownoutdoorenvironmentsrealtimeflightexperiments
AT deokjinlee deepreinforcementlearningforendtoendlocalmotionplanningofautonomousaerialrobotsinunknownoutdoorenvironmentsrealtimeflightexperiments