Teaching Humanoid Robot Reaching Motion by Imitation and Reinforcement Learning

被引:1
|
作者
Savevska, Kristina [1 ,2 ]
Ude, Ales [1 ]
机构
[1] Jozef Stefan Inst, Dept Automat Biocybernet & Robot, Humanoid & Cognit Robot Lab, Jamova Cesta 39, Ljubljana 1000, Slovenia
[2] Int Postgrad Sch Jozef Stefan, Jamova Cesta 39, Ljubljana 1000, Slovenia
关键词
Humanoids; Imitation Learning; Reinforcement learning;
D O I
10.1007/978-3-031-32606-6_7
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper presents a user-friendly method for programming humanoid robots without the need for expert knowledge. We propose a combination of imitation learning and reinforcement learning to teach and optimize demonstrated trajectories. An initial trajectory for reinforcement learning is generated using a stable whole-body motion imitation system. The acquired motion is then refined using a stochastic optimal control-based reinforcement learning algorithm called Policy Improvement with Path Integrals with Covariance Matrix Adaptation (PI2-CMA). We tested the approach for programming humanoid robot reaching motion. Our experimental results show that the proposed approach is successful at learning reaching motions while preserving the postural balance of the robot. We also show how a stable humanoid robot trajectory learned in simulation can be effectively adapted to different dynamic environments, e.g. a different simulator or a real robot. The resulting learning methodology allows for quick and efficient optimization of the demonstrated trajectories while also taking into account the constraints of the desired task. The learning methodology was tested in a simulated environment and on the real humanoid robot TALOS.
引用
收藏
页码:53 / 61
页数:9
相关论文
共 50 条
  • [21] Deep Reinforcement Learning for Humanoid Robot Behaviors
    Muzio, Alexandre F. V.
    Maximo, Marcos R. O. A.
    Yoneyama, Takashi
    JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 2022, 105 (01)
  • [22] Humanoid robot control based on reinforcement learning
    Iida, S
    Kuwayama, K
    Kanoh, M
    Kato, S
    Kunitachi, T
    Itoh, H
    PROCEEDINGS OF THE 2004 INTERNATIONAL SYMPOSIUM ON MICRO-NANOMECHATRONICS AND HUMAN SCIENCE, 2004, : 353 - 358
  • [23] Humanoid robot control based on reinforcement learning
    Iida, S. (iida@ics.nitech.ac.jp), IEEE Robotics and Automation Society; Nagoya University, Japan; City of Nagoya, Japan; Nagoya City Science Museum; Chubu Science and Technology Center (Institute of Electrical and Electronics Engineers Inc.):
  • [24] A Reinforcement Learning Method for Humanoid Robot Walking
    Liu, Yunda
    Bi, Sheng
    Dong, Min
    Zhang, Yingjie
    Huang, Jialing
    Zhang, Jiawei
    2018 IEEE 8TH ANNUAL INTERNATIONAL CONFERENCE ON CYBER TECHNOLOGY IN AUTOMATION, CONTROL, AND INTELLIGENT SYSTEMS (IEEE-CYBER), 2018, : 623 - 628
  • [25] Deep Reinforcement Learning for Humanoid Robot Dribbling
    Muzio, Alexandre F., V
    Maximo, Marcos R. O. A.
    Yoneyama, Takashi
    2020 XVIII LATIN AMERICAN ROBOTICS SYMPOSIUM, 2020 XII BRAZILIAN SYMPOSIUM ON ROBOTICS AND 2020 XI WORKSHOP OF ROBOTICS IN EDUCATION (LARS-SBR-WRE 2020), 2020, : 246 - 251
  • [26] Deep Reinforcement Learning for Humanoid Robot Behaviors
    Alexandre F. V. Muzio
    Marcos R. O. A. Maximo
    Takashi Yoneyama
    Journal of Intelligent & Robotic Systems, 2022, 105
  • [27] Deep Reinforcement Learning for Humanoid Robot Behaviors
    Muzio, Alexandre F. V.
    Maximo, Marcos R. O. A.
    Yoneyama, Takashi
    Journal of Intelligent and Robotic Systems: Theory and Applications, 2022, 105 (01):
  • [28] Humanoid Robot Gait Imitation
    Hwang, Kao-Shing
    Lin, Jin-Ling
    Huang, Tsung-Chuan
    Hsu, Hsin-Jung
    2014 PROCEEDINGS OF THE SICE ANNUAL CONFERENCE (SICE), 2014, : 2124 - +
  • [29] Prediction-Based Synchronized Human Motion Imitation by a Humanoid Robot
    Hu, Kai
    Lee, Dongheui
    AT-AUTOMATISIERUNGSTECHNIK, 2012, 60 (11) : 705 - 714
  • [30] Offline Imitation of a Human Motion by a Humanoid Robot Under Balance Constraint
    Munirathinam, K.
    Chevallereau, C.
    Sakka, S.
    NEW TRENDS IN MEDICAL AND SERVICE ROBOTS: CHALLENGES AND SOLUTIONS, 2014, 20 : 269 - 282