A Self-Adaptive Double Q-Backstepping Trajectory Tracking Control Approach Based on Reinforcement Learning for Mobile Robots

When a mobile robot inspects tasks with complex requirements indoors, the traditional backstepping method cannot guarantee the accuracy of the trajectory, leading to problems such as the instrument not being inside the image and focus failure when the robot grabs the image with high zoom. In order t...

Full description

Bibliographic Details
Main Authors: Naifeng He, Zhong Yang, Xiaoliang Fan, Jiying Wu, Yaoyu Sui, Qiuyan Zhang
Format: Article
Language:English
Published: MDPI AG 2023-08-01
Series:Actuators
Subjects:
Online Access:https://www.mdpi.com/2076-0825/12/8/326
_version_ 1797585868879298560
author Naifeng He
Zhong Yang
Xiaoliang Fan
Jiying Wu
Yaoyu Sui
Qiuyan Zhang
author_facet Naifeng He
Zhong Yang
Xiaoliang Fan
Jiying Wu
Yaoyu Sui
Qiuyan Zhang
author_sort Naifeng He
collection DOAJ
description When a mobile robot inspects tasks with complex requirements indoors, the traditional backstepping method cannot guarantee the accuracy of the trajectory, leading to problems such as the instrument not being inside the image and focus failure when the robot grabs the image with high zoom. In order to solve this problem, this paper proposes an adaptive backstepping method based on double Q-learning for tracking and controlling the trajectory of mobile robots. We design the incremental model-free algorithm of Double-Q learning, which can quickly learn to rectify the trajectory tracking controller gain online. For the controller gain rectification problem in non-uniform state space exploration, we propose an incremental active learning exploration algorithm that incorporates memory playback as well as experience playback mechanisms to achieve online fast learning and controller gain rectification for agents. To verify the feasibility of the algorithm, we perform algorithm verification on different types of trajectories in Gazebo and physical platforms. The results show that the adaptive trajectory tracking control algorithm can be used to rectify the mobile robot trajectory tracking controller’s gain. Compared with the Backstepping-Fractional-Older PID controller and Fuzzy-Backstepping controller, Double Q-backstepping has better robustness, generalization, real-time, and stronger anti-disturbance capability.
first_indexed 2024-03-11T00:13:15Z
format Article
id doaj.art-2f0141187cf24d8ebc0501dbe90e3cb5
institution Directory Open Access Journal
issn 2076-0825
language English
last_indexed 2024-03-11T00:13:15Z
publishDate 2023-08-01
publisher MDPI AG
record_format Article
series Actuators
spelling doaj.art-2f0141187cf24d8ebc0501dbe90e3cb52023-11-18T23:49:01ZengMDPI AGActuators2076-08252023-08-0112832610.3390/act12080326A Self-Adaptive Double Q-Backstepping Trajectory Tracking Control Approach Based on Reinforcement Learning for Mobile RobotsNaifeng He0Zhong Yang1Xiaoliang Fan2Jiying Wu3Yaoyu Sui4Qiuyan Zhang5College of Automation Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing 211106, ChinaCollege of Automation Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing 211106, ChinaState Key Laboratory of Robotics, Shenyang Institute of Automation Chinese Academy of Sciences, Shenyang 110017, ChinaCollege of Automation Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing 211106, ChinaCollege of Automation Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing 211106, ChinaElectric Power Research Institute of Guizhou Power Grid Co., Ltd., Guiyang 550002, ChinaWhen a mobile robot inspects tasks with complex requirements indoors, the traditional backstepping method cannot guarantee the accuracy of the trajectory, leading to problems such as the instrument not being inside the image and focus failure when the robot grabs the image with high zoom. In order to solve this problem, this paper proposes an adaptive backstepping method based on double Q-learning for tracking and controlling the trajectory of mobile robots. We design the incremental model-free algorithm of Double-Q learning, which can quickly learn to rectify the trajectory tracking controller gain online. For the controller gain rectification problem in non-uniform state space exploration, we propose an incremental active learning exploration algorithm that incorporates memory playback as well as experience playback mechanisms to achieve online fast learning and controller gain rectification for agents. To verify the feasibility of the algorithm, we perform algorithm verification on different types of trajectories in Gazebo and physical platforms. The results show that the adaptive trajectory tracking control algorithm can be used to rectify the mobile robot trajectory tracking controller’s gain. Compared with the Backstepping-Fractional-Older PID controller and Fuzzy-Backstepping controller, Double Q-backstepping has better robustness, generalization, real-time, and stronger anti-disturbance capability.https://www.mdpi.com/2076-0825/12/8/326reinforcement learningdouble Q-backstepping controlmobile robottrajectory tracking control
spellingShingle Naifeng He
Zhong Yang
Xiaoliang Fan
Jiying Wu
Yaoyu Sui
Qiuyan Zhang
A Self-Adaptive Double Q-Backstepping Trajectory Tracking Control Approach Based on Reinforcement Learning for Mobile Robots
Actuators
reinforcement learning
double Q-backstepping control
mobile robot
trajectory tracking control
title A Self-Adaptive Double Q-Backstepping Trajectory Tracking Control Approach Based on Reinforcement Learning for Mobile Robots
title_full A Self-Adaptive Double Q-Backstepping Trajectory Tracking Control Approach Based on Reinforcement Learning for Mobile Robots
title_fullStr A Self-Adaptive Double Q-Backstepping Trajectory Tracking Control Approach Based on Reinforcement Learning for Mobile Robots
title_full_unstemmed A Self-Adaptive Double Q-Backstepping Trajectory Tracking Control Approach Based on Reinforcement Learning for Mobile Robots
title_short A Self-Adaptive Double Q-Backstepping Trajectory Tracking Control Approach Based on Reinforcement Learning for Mobile Robots
title_sort self adaptive double q backstepping trajectory tracking control approach based on reinforcement learning for mobile robots
topic reinforcement learning
double Q-backstepping control
mobile robot
trajectory tracking control
url https://www.mdpi.com/2076-0825/12/8/326
work_keys_str_mv AT naifenghe aselfadaptivedoubleqbacksteppingtrajectorytrackingcontrolapproachbasedonreinforcementlearningformobilerobots
AT zhongyang aselfadaptivedoubleqbacksteppingtrajectorytrackingcontrolapproachbasedonreinforcementlearningformobilerobots
AT xiaoliangfan aselfadaptivedoubleqbacksteppingtrajectorytrackingcontrolapproachbasedonreinforcementlearningformobilerobots
AT jiyingwu aselfadaptivedoubleqbacksteppingtrajectorytrackingcontrolapproachbasedonreinforcementlearningformobilerobots
AT yaoyusui aselfadaptivedoubleqbacksteppingtrajectorytrackingcontrolapproachbasedonreinforcementlearningformobilerobots
AT qiuyanzhang aselfadaptivedoubleqbacksteppingtrajectorytrackingcontrolapproachbasedonreinforcementlearningformobilerobots
AT naifenghe selfadaptivedoubleqbacksteppingtrajectorytrackingcontrolapproachbasedonreinforcementlearningformobilerobots
AT zhongyang selfadaptivedoubleqbacksteppingtrajectorytrackingcontrolapproachbasedonreinforcementlearningformobilerobots
AT xiaoliangfan selfadaptivedoubleqbacksteppingtrajectorytrackingcontrolapproachbasedonreinforcementlearningformobilerobots
AT jiyingwu selfadaptivedoubleqbacksteppingtrajectorytrackingcontrolapproachbasedonreinforcementlearningformobilerobots
AT yaoyusui selfadaptivedoubleqbacksteppingtrajectorytrackingcontrolapproachbasedonreinforcementlearningformobilerobots
AT qiuyanzhang selfadaptivedoubleqbacksteppingtrajectorytrackingcontrolapproachbasedonreinforcementlearningformobilerobots