Ensemble Bootstrapped Deep Deterministic Policy Gradient for Vision-Based Robotic Grasping
With sufficient practice, humans can grab objects they have never seen before through brain decision-making. However, the manipulators, which has a wide range of applications in industrial production, can still only grab specific objects. Because most of the grasp algorithms rely on prior knowledge...
Main Authors: | , , , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2021-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/9316755/ |
_version_ | 1818736208698671104 |
---|---|
author | Weiwei Liu Linpeng Peng Junjie Cao Xiaokuan Fu Yong Liu Zaisheng Pan Jian Yang |
author_facet | Weiwei Liu Linpeng Peng Junjie Cao Xiaokuan Fu Yong Liu Zaisheng Pan Jian Yang |
author_sort | Weiwei Liu |
collection | DOAJ |
description | With sufficient practice, humans can grab objects they have never seen before through brain decision-making. However, the manipulators, which has a wide range of applications in industrial production, can still only grab specific objects. Because most of the grasp algorithms rely on prior knowledge such as hand-eye calibration results, object model features, and can only target specific types of objects. When the task scenario and the operation target change, it cannot perform effective redeployment. In order to solve the above problems, academia often uses reinforcement learning to train grasping algorithms. However, the method of reinforcement learning in the field of manipulators grasping mainly encounters these main problems: insufficient sample utilization, poor algorithm stability, and limited exploration. This article uses LfD, BC, and DDPG to improve sample utilization. Use multiple critics to integrate and evaluate input actions to solve the problem of algorithm instability. Finally, inspired by Thompson's sampling idea, the input action is evaluated from different angles, which increases the algorithm's exploration of the environment and reduces the number of interactions with the environment. EDDPG and EBDDPG algorithm is designed in the article. In order to further improve the generalization ability of the algorithm, this article does not use extra information that is difficult to obtain directly on the physical platform, such as the real coordinates of the target object and the continuous motion space at the end of the manipulator in the Cartesian coordinate system is used as the output of the decision. The simulation results show that, under the same number of interactions, the manipulators' success rate in grabbing 1000 random objects has increased more than double and reached state-of-the-art(SOTA) performance. |
first_indexed | 2024-12-18T00:33:30Z |
format | Article |
id | doaj.art-ecc8057156e74493ab539be9a9d6d660 |
institution | Directory Open Access Journal |
issn | 2169-3536 |
language | English |
last_indexed | 2024-12-18T00:33:30Z |
publishDate | 2021-01-01 |
publisher | IEEE |
record_format | Article |
series | IEEE Access |
spelling | doaj.art-ecc8057156e74493ab539be9a9d6d6602022-12-21T21:27:04ZengIEEEIEEE Access2169-35362021-01-019199161992510.1109/ACCESS.2021.30498609316755Ensemble Bootstrapped Deep Deterministic Policy Gradient for Vision-Based Robotic GraspingWeiwei Liu0https://orcid.org/0000-0002-2496-7748Linpeng Peng1https://orcid.org/0000-0003-1754-9879Junjie Cao2Xiaokuan Fu3https://orcid.org/0000-0001-7115-5091Yong Liu4https://orcid.org/0000-0003-4822-8939Zaisheng Pan5Jian Yang6Institute of Cyber-Systems and Control, Zhejiang University, Hangzhou, ChinaInstitute of Cyber-Systems and Control, Zhejiang University, Hangzhou, ChinaInstitute of Cyber-Systems and Control, Zhejiang University, Hangzhou, ChinaInstitute of Cyber-Systems and Control, Zhejiang University, Hangzhou, ChinaInstitute of Cyber-Systems and Control, Zhejiang University, Hangzhou, ChinaInstitute of Cyber-Systems and Control, Zhejiang University, Hangzhou, ChinaChina Research and Development Academy of Machinery Equipment, Beijing, ChinaWith sufficient practice, humans can grab objects they have never seen before through brain decision-making. However, the manipulators, which has a wide range of applications in industrial production, can still only grab specific objects. Because most of the grasp algorithms rely on prior knowledge such as hand-eye calibration results, object model features, and can only target specific types of objects. When the task scenario and the operation target change, it cannot perform effective redeployment. In order to solve the above problems, academia often uses reinforcement learning to train grasping algorithms. However, the method of reinforcement learning in the field of manipulators grasping mainly encounters these main problems: insufficient sample utilization, poor algorithm stability, and limited exploration. This article uses LfD, BC, and DDPG to improve sample utilization. Use multiple critics to integrate and evaluate input actions to solve the problem of algorithm instability. Finally, inspired by Thompson's sampling idea, the input action is evaluated from different angles, which increases the algorithm's exploration of the environment and reduces the number of interactions with the environment. EDDPG and EBDDPG algorithm is designed in the article. In order to further improve the generalization ability of the algorithm, this article does not use extra information that is difficult to obtain directly on the physical platform, such as the real coordinates of the target object and the continuous motion space at the end of the manipulator in the Cartesian coordinate system is used as the output of the decision. The simulation results show that, under the same number of interactions, the manipulators' success rate in grabbing 1000 random objects has increased more than double and reached state-of-the-art(SOTA) performance.https://ieeexplore.ieee.org/document/9316755/Robotic graspingreinforcement learningensemble learningThompson sampling |
spellingShingle | Weiwei Liu Linpeng Peng Junjie Cao Xiaokuan Fu Yong Liu Zaisheng Pan Jian Yang Ensemble Bootstrapped Deep Deterministic Policy Gradient for Vision-Based Robotic Grasping IEEE Access Robotic grasping reinforcement learning ensemble learning Thompson sampling |
title | Ensemble Bootstrapped Deep Deterministic Policy Gradient for Vision-Based Robotic Grasping |
title_full | Ensemble Bootstrapped Deep Deterministic Policy Gradient for Vision-Based Robotic Grasping |
title_fullStr | Ensemble Bootstrapped Deep Deterministic Policy Gradient for Vision-Based Robotic Grasping |
title_full_unstemmed | Ensemble Bootstrapped Deep Deterministic Policy Gradient for Vision-Based Robotic Grasping |
title_short | Ensemble Bootstrapped Deep Deterministic Policy Gradient for Vision-Based Robotic Grasping |
title_sort | ensemble bootstrapped deep deterministic policy gradient for vision based robotic grasping |
topic | Robotic grasping reinforcement learning ensemble learning Thompson sampling |
url | https://ieeexplore.ieee.org/document/9316755/ |
work_keys_str_mv | AT weiweiliu ensemblebootstrappeddeepdeterministicpolicygradientforvisionbasedroboticgrasping AT linpengpeng ensemblebootstrappeddeepdeterministicpolicygradientforvisionbasedroboticgrasping AT junjiecao ensemblebootstrappeddeepdeterministicpolicygradientforvisionbasedroboticgrasping AT xiaokuanfu ensemblebootstrappeddeepdeterministicpolicygradientforvisionbasedroboticgrasping AT yongliu ensemblebootstrappeddeepdeterministicpolicygradientforvisionbasedroboticgrasping AT zaishengpan ensemblebootstrappeddeepdeterministicpolicygradientforvisionbasedroboticgrasping AT jianyang ensemblebootstrappeddeepdeterministicpolicygradientforvisionbasedroboticgrasping |