A Deep Reinforcement-Learning Approach for Inverse Kinematics Solution of a High Degree of Freedom Robotic Manipulator

The foundation and emphasis of robotic manipulator control is Inverse Kinematics (IK). Due to the complexity of derivation, difficulty of computation, and redundancy, traditional IK solutions pose numerous challenges to the operation of a variety of robotic manipulators. This paper develops a Deep R...

Full description

Bibliographic Details
Main Authors: Aryslan Malik, Yevgeniy Lischuk, Troy Henderson, Richard Prazenica
Format: Article
Language:English
Published: MDPI AG 2022-04-01
Series:Robotics
Subjects:
Online Access:https://www.mdpi.com/2218-6581/11/2/44
_version_ 1797434252164333568
author Aryslan Malik
Yevgeniy Lischuk
Troy Henderson
Richard Prazenica
author_facet Aryslan Malik
Yevgeniy Lischuk
Troy Henderson
Richard Prazenica
author_sort Aryslan Malik
collection DOAJ
description The foundation and emphasis of robotic manipulator control is Inverse Kinematics (IK). Due to the complexity of derivation, difficulty of computation, and redundancy, traditional IK solutions pose numerous challenges to the operation of a variety of robotic manipulators. This paper develops a Deep Reinforcement Learning (RL) approach for solving the IK problem of a 7-Degree of Freedom (DOF) robotic manipulator using Product of Exponentials (PoE) as a Forward Kinematics (FK) computation tool and the Deep Q-Network (DQN) as an IK solver. The selected approach is architecturally simpler, making it faster and easier to implement, as well as more stable, because it is less sensitive to hyperparameters than continuous action spaces algorithms. The algorithm is designed to produce joint-space trajectories from a given end-effector trajectory. Different network architectures were explored and the output of the DQN was implemented experimentally on a Sawyer robotic arm. The DQN was able to find different trajectories corresponding to a specified Cartesian path of the end-effector. The network agent was able to learn random Bézier and straight-line end-effector trajectories in a reasonable time frame with good accuracy, demonstrating that even though DQN is mainly used in discrete solution spaces, it could be applied to generate joint space trajectories.
first_indexed 2024-03-09T10:30:31Z
format Article
id doaj.art-d8213e1d3313498caac5b4457ec63798
institution Directory Open Access Journal
issn 2218-6581
language English
last_indexed 2024-03-09T10:30:31Z
publishDate 2022-04-01
publisher MDPI AG
record_format Article
series Robotics
spelling doaj.art-d8213e1d3313498caac5b4457ec637982023-12-01T21:22:17ZengMDPI AGRobotics2218-65812022-04-011124410.3390/robotics11020044A Deep Reinforcement-Learning Approach for Inverse Kinematics Solution of a High Degree of Freedom Robotic ManipulatorAryslan Malik0Yevgeniy Lischuk1Troy Henderson2Richard Prazenica3Aerospace Engineering Department, Embry—Riddle Aeronautical University, Daytona Beach, FL 32114, USASoftware—Device OS, Amazon, Austin, TX 78758, USAAerospace Engineering Department, Embry—Riddle Aeronautical University, Daytona Beach, FL 32114, USAAerospace Engineering Department, Embry—Riddle Aeronautical University, Daytona Beach, FL 32114, USAThe foundation and emphasis of robotic manipulator control is Inverse Kinematics (IK). Due to the complexity of derivation, difficulty of computation, and redundancy, traditional IK solutions pose numerous challenges to the operation of a variety of robotic manipulators. This paper develops a Deep Reinforcement Learning (RL) approach for solving the IK problem of a 7-Degree of Freedom (DOF) robotic manipulator using Product of Exponentials (PoE) as a Forward Kinematics (FK) computation tool and the Deep Q-Network (DQN) as an IK solver. The selected approach is architecturally simpler, making it faster and easier to implement, as well as more stable, because it is less sensitive to hyperparameters than continuous action spaces algorithms. The algorithm is designed to produce joint-space trajectories from a given end-effector trajectory. Different network architectures were explored and the output of the DQN was implemented experimentally on a Sawyer robotic arm. The DQN was able to find different trajectories corresponding to a specified Cartesian path of the end-effector. The network agent was able to learn random Bézier and straight-line end-effector trajectories in a reasonable time frame with good accuracy, demonstrating that even though DQN is mainly used in discrete solution spaces, it could be applied to generate joint space trajectories.https://www.mdpi.com/2218-6581/11/2/44roboticsrobotic manipulatorInverse Kinematicsneural networksReinforcement LearningDeep Learning
spellingShingle Aryslan Malik
Yevgeniy Lischuk
Troy Henderson
Richard Prazenica
A Deep Reinforcement-Learning Approach for Inverse Kinematics Solution of a High Degree of Freedom Robotic Manipulator
Robotics
robotics
robotic manipulator
Inverse Kinematics
neural networks
Reinforcement Learning
Deep Learning
title A Deep Reinforcement-Learning Approach for Inverse Kinematics Solution of a High Degree of Freedom Robotic Manipulator
title_full A Deep Reinforcement-Learning Approach for Inverse Kinematics Solution of a High Degree of Freedom Robotic Manipulator
title_fullStr A Deep Reinforcement-Learning Approach for Inverse Kinematics Solution of a High Degree of Freedom Robotic Manipulator
title_full_unstemmed A Deep Reinforcement-Learning Approach for Inverse Kinematics Solution of a High Degree of Freedom Robotic Manipulator
title_short A Deep Reinforcement-Learning Approach for Inverse Kinematics Solution of a High Degree of Freedom Robotic Manipulator
title_sort deep reinforcement learning approach for inverse kinematics solution of a high degree of freedom robotic manipulator
topic robotics
robotic manipulator
Inverse Kinematics
neural networks
Reinforcement Learning
Deep Learning
url https://www.mdpi.com/2218-6581/11/2/44
work_keys_str_mv AT aryslanmalik adeepreinforcementlearningapproachforinversekinematicssolutionofahighdegreeoffreedomroboticmanipulator
AT yevgeniylischuk adeepreinforcementlearningapproachforinversekinematicssolutionofahighdegreeoffreedomroboticmanipulator
AT troyhenderson adeepreinforcementlearningapproachforinversekinematicssolutionofahighdegreeoffreedomroboticmanipulator
AT richardprazenica adeepreinforcementlearningapproachforinversekinematicssolutionofahighdegreeoffreedomroboticmanipulator
AT aryslanmalik deepreinforcementlearningapproachforinversekinematicssolutionofahighdegreeoffreedomroboticmanipulator
AT yevgeniylischuk deepreinforcementlearningapproachforinversekinematicssolutionofahighdegreeoffreedomroboticmanipulator
AT troyhenderson deepreinforcementlearningapproachforinversekinematicssolutionofahighdegreeoffreedomroboticmanipulator
AT richardprazenica deepreinforcementlearningapproachforinversekinematicssolutionofahighdegreeoffreedomroboticmanipulator