Using Implicit Behavior Cloning and Dynamic Movement Primitive to Facilitate Reinforcement Learning for Robot Motion Planning

被引:0
|
作者
Zhang, Zengjie [1 ]
Hong, Jayden [2 ]
Enayati, Amir M. Soufi [2 ]
Najjaran, Homayoun [2 ]
机构
[1] Eindhoven Univ Technol, Dept Elect Engn, NL-5612 AZ Eindhoven, Netherlands
[2] Univ Victoria, Fac Engn & Comp Sci, V8P 5C2 Victoria, BC, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Planning; Training; Robots; Trajectory; Robot motion; Force; Dynamics; Behavior cloning (BC); heuristic method; human motion; learning from demonstration; motion primitive; reinforcement learning (RL); robot motion planning; END-TO-END; OPTIMIZATION;
D O I
10.1109/TRO.2024.3468770
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Reinforcement learning (RL) for motion planning of multi-degree-of-freedom robots still suffers from low efficiency in terms of slow training speed and poor generalizability. In this article, we propose a novel RL-based robot motion planning framework that uses implicit behavior cloning (IBC) and dynamic movement primitive (DMP) to improve the training speed and generalizability of an off-policy RL agent. IBC utilizes human demonstration data to leverage the training speed of RL, and DMP serves as a heuristic model that transfers motion planning into a simpler planning space. To support this, we also create a human demonstration dataset using a pick-and-place experiment that can be used for similar studies. Comparison studies reveal the advantage of the proposed method over the conventional RL agents with faster training speed and higher scores. A real-robot experiment indicates the applicability of the proposed method to a simple assembly task. Our work provides a novel perspective on using motion primitives and human demonstration to leverage the performance of RL for robot applications.
引用
收藏
页码:4733 / 4749
页数:17
相关论文
共 50 条
  • [1] Optimal Robot Motion Planning in Constrained Workspaces Using Reinforcement Learning
    Rousseas, Panagiotis
    Bechlioulis, Charalampos P.
    Kyriakopoulos, Kostas J.
    2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 6917 - 6922
  • [2] The Arm Planning with Dynamic Movement Primitive for Humanoid Service Robot
    Lin, Menglei
    Lu, Zhiguo
    Wang, Shixiong
    Wang, Ruchao
    2020 5TH INTERNATIONAL CONFERENCE ON ADVANCED ROBOTICS AND MECHATRONICS (ICARM 2020), 2020, : 513 - 518
  • [3] Motion generation for walking exoskeleton robot using multiple dynamic movement primitives sequences combined with reinforcement learning
    Zhang, Peng
    Zhang, Junxia
    ROBOTICA, 2022, 40 (08) : 2732 - 2747
  • [4] Robot Motion Planning Under Uncertain Condition Using Deep Reinforcement Learning
    Chen, Zhuang
    Zhou, Lin
    Guo, Min
    PROCEEDINGS OF THE 2018 INTERNATIONAL CONFERENCE ON MECHANICAL, ELECTRONIC, CONTROL AND AUTOMATION ENGINEERING (MECAE 2018), 2018, 149 : 94 - 100
  • [5] Trajectory generation using reinforcement learning for autonomous helicopter with adaptive dynamic movement primitive
    Guo, Xiao
    PROCEEDINGS OF THE INSTITUTION OF MECHANICAL ENGINEERS PART I-JOURNAL OF SYSTEMS AND CONTROL ENGINEERING, 2017, 231 (06) : 495 - 509
  • [6] Robot Learning from Multiple Demonstrations with Dynamic Movement Primitive
    Chen, Chuize
    Yang, Chenguang
    Zeng, Chao
    Wang, Ning
    Li, Zhijun
    2017 2ND INTERNATIONAL CONFERENCE ON ADVANCED ROBOTICS AND MECHATRONICS (ICARM), 2017, : 523 - 528
  • [7] Hierarchical dynamic movement primitive for the smooth movement of robots based on deep reinforcement learning
    Yuan, Yinlong
    Yu, Zhu Liang
    Hua, Liang
    Cheng, Yun
    Li, Junhong
    Sang, Xiaohu
    APPLIED INTELLIGENCE, 2023, 53 (02) : 1417 - 1434
  • [8] Hierarchical dynamic movement primitive for the smooth movement of robots based on deep reinforcement learning
    Yinlong Yuan
    Zhu Liang Yu
    Liang Hua
    Yun Cheng
    Junhong Li
    Xiaohu Sang
    Applied Intelligence, 2023, 53 : 1417 - 1434
  • [9] Motion Planning and Control with Randomized Payloads on Real Robot Using Deep Reinforcement Learning
    Demir, Ali
    Sezer, Volkan
    INTERNATIONAL JOURNAL OF SEMANTIC COMPUTING, 2019, 13 (04) : 541 - 563
  • [10] Robot path planning in dynamic environment based on reinforcement learning
    Zhuang, Xiao-Dong
    Meng, Qing-Chun
    Wei, Tian-Bin
    Wang, Xu-Zhu
    Tan, Rui
    Li, Xiao-Jing
    Journal of Harbin Institute of Technology (New Series), 2001, 8 (03) : 253 - 255