Reinforcement Learning with Dynamic Movement Primitives for Obstacle Avoidance

Dynamic movement primitives (DMPs) are a robust framework for movement generation from demonstrations. This framework can be extended by adding a perturbing term to achieve obstacle avoidance without sacrificing stability. The additional term is usually constructed based on potential functions. Alth...

Full description

Bibliographic Details
Main Authors: Ang Li, Zhenze Liu, Wenrui Wang, Mingchao Zhu, Yanhui Li, Qi Huo, Ming Dai
Format: Article
Language:English
Published: MDPI AG 2021-11-01
Series:Applied Sciences
Subjects:
Online Access:https://www.mdpi.com/2076-3417/11/23/11184
_version_ 1797508069510348800
author Ang Li
Zhenze Liu
Wenrui Wang
Mingchao Zhu
Yanhui Li
Qi Huo
Ming Dai
author_facet Ang Li
Zhenze Liu
Wenrui Wang
Mingchao Zhu
Yanhui Li
Qi Huo
Ming Dai
author_sort Ang Li
collection DOAJ
description Dynamic movement primitives (DMPs) are a robust framework for movement generation from demonstrations. This framework can be extended by adding a perturbing term to achieve obstacle avoidance without sacrificing stability. The additional term is usually constructed based on potential functions. Although different potentials are adopted to improve the performance of obstacle avoidance, the profiles of potentials are rarely incorporated into reinforcement learning (RL) framework. In this contribution, we present a RL based method to learn not only the profiles of potentials but also the shape parameters of a motion. The algorithm employed is PI2 (Policy Improvement with Path Integrals), a model-free, sampling-based learning method. By using the PI2, the profiles of potentials and the parameters of the DMPs are learned simultaneously; therefore, we can optimize obstacle avoidance while completing specified tasks. We validate the presented method in simulations and with a redundant robot arm in experiments.
first_indexed 2024-03-10T04:58:02Z
format Article
id doaj.art-4821f3a499de4af392173b06decd54ef
institution Directory Open Access Journal
issn 2076-3417
language English
last_indexed 2024-03-10T04:58:02Z
publishDate 2021-11-01
publisher MDPI AG
record_format Article
series Applied Sciences
spelling doaj.art-4821f3a499de4af392173b06decd54ef2023-11-23T02:03:51ZengMDPI AGApplied Sciences2076-34172021-11-0111231118410.3390/app112311184Reinforcement Learning with Dynamic Movement Primitives for Obstacle AvoidanceAng Li0Zhenze Liu1Wenrui Wang2Mingchao Zhu3Yanhui Li4Qi Huo5Ming Dai6Changchun Institute of Optics, Fine Mechanics and Physics, Chinese Academy of Sciences, Changchun 130033, ChinaCollege of Communication Engineering, Jilin University, Changchun 130025, ChinaChangchun Institute of Optics, Fine Mechanics and Physics, Chinese Academy of Sciences, Changchun 130033, ChinaChangchun Institute of Optics, Fine Mechanics and Physics, Chinese Academy of Sciences, Changchun 130033, ChinaChangchun Institute of Optics, Fine Mechanics and Physics, Chinese Academy of Sciences, Changchun 130033, ChinaChangchun Institute of Optics, Fine Mechanics and Physics, Chinese Academy of Sciences, Changchun 130033, ChinaChangchun Institute of Optics, Fine Mechanics and Physics, Chinese Academy of Sciences, Changchun 130033, ChinaDynamic movement primitives (DMPs) are a robust framework for movement generation from demonstrations. This framework can be extended by adding a perturbing term to achieve obstacle avoidance without sacrificing stability. The additional term is usually constructed based on potential functions. Although different potentials are adopted to improve the performance of obstacle avoidance, the profiles of potentials are rarely incorporated into reinforcement learning (RL) framework. In this contribution, we present a RL based method to learn not only the profiles of potentials but also the shape parameters of a motion. The algorithm employed is PI2 (Policy Improvement with Path Integrals), a model-free, sampling-based learning method. By using the PI2, the profiles of potentials and the parameters of the DMPs are learned simultaneously; therefore, we can optimize obstacle avoidance while completing specified tasks. We validate the presented method in simulations and with a redundant robot arm in experiments.https://www.mdpi.com/2076-3417/11/23/11184obstacle avoidanceDynamic Movement Primitivesreinforcement learningPI2 (policy improvement with path integrals)
spellingShingle Ang Li
Zhenze Liu
Wenrui Wang
Mingchao Zhu
Yanhui Li
Qi Huo
Ming Dai
Reinforcement Learning with Dynamic Movement Primitives for Obstacle Avoidance
Applied Sciences
obstacle avoidance
Dynamic Movement Primitives
reinforcement learning
PI2 (policy improvement with path integrals)
title Reinforcement Learning with Dynamic Movement Primitives for Obstacle Avoidance
title_full Reinforcement Learning with Dynamic Movement Primitives for Obstacle Avoidance
title_fullStr Reinforcement Learning with Dynamic Movement Primitives for Obstacle Avoidance
title_full_unstemmed Reinforcement Learning with Dynamic Movement Primitives for Obstacle Avoidance
title_short Reinforcement Learning with Dynamic Movement Primitives for Obstacle Avoidance
title_sort reinforcement learning with dynamic movement primitives for obstacle avoidance
topic obstacle avoidance
Dynamic Movement Primitives
reinforcement learning
PI2 (policy improvement with path integrals)
url https://www.mdpi.com/2076-3417/11/23/11184
work_keys_str_mv AT angli reinforcementlearningwithdynamicmovementprimitivesforobstacleavoidance
AT zhenzeliu reinforcementlearningwithdynamicmovementprimitivesforobstacleavoidance
AT wenruiwang reinforcementlearningwithdynamicmovementprimitivesforobstacleavoidance
AT mingchaozhu reinforcementlearningwithdynamicmovementprimitivesforobstacleavoidance
AT yanhuili reinforcementlearningwithdynamicmovementprimitivesforobstacleavoidance
AT qihuo reinforcementlearningwithdynamicmovementprimitivesforobstacleavoidance
AT mingdai reinforcementlearningwithdynamicmovementprimitivesforobstacleavoidance