Reinforcement learning control for a three-link biped robot with energy-efficient periodic gaits

被引:2
|
作者
Pan, Zebang [1 ]
Yin, Shan [1 ]
Wen, Guilin [2 ]
Tan, Zhao [1 ]
机构
[1] Hunan Univ, State Key Lab Adv Design & Manufacture Vehicle Bod, Changsha 410082, Peoples R China
[2] Yanshan Univ, Sch Mech Engn, Qinhuangdao 066004, Peoples R China
基金
中国国家自然科学基金; 中国博士后科学基金;
关键词
Three-link biped robot; Deep Reinforcement learning; Periodic gaits; Energy optimization; STABLE WALKING; LOCOMOTION; COST; FEET;
D O I
10.1007/s10409-022-22304-x
中图分类号
TH [机械、仪表工业];
学科分类号
0802 ;
摘要
Designing a high-performance controller for the walking gaits of biped robots remains an open research area due to their strong nonlinearity and non-smooth responses. To overcome such challenges, a humanoid robot with a torso, i.e., a three-link biped robot involving both impact and friction, is developed firstly. Then, the twin delayed deep deterministic policy gradient algorithm is adopted to design the reinforcement learning controller for the proposed biped robot. For the specified control targets, i.e., energy-efficient periodic gaits for both the downhill and uphill cases, a reward function utilizing the Poincare map and the power function is constructed to provide guidelines for the controller. Thus, the proposed controller can learn to adaptively output accurate cosine torques to achieve the goal without relying on the pre-designed reference trajectories or embedded unstable periodic gaits. A comparative study between the proposed reinforcement learning and neural network proportion differentiation controllers demonstrates the proposed controller can lead to accurate and energy-efficient periodic gaits and provide strong adaptability and robustness within a wide variety of walking slopes.
引用
收藏
页数:23
相关论文
共 50 条
  • [31] Deep Reinforcement Learning for Energy-Efficient Power Control in Heterogeneous Networks
    Peng, Jianhao
    Zheng, Jiabao
    Zhang, Lin
    Xiao, Ming
    IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC 2022), 2022, : 141 - 146
  • [32] Standing balance control of humanoid robot based on three-link dynamic model
    Zhu Q.
    Wu H.
    Wu J.
    Xiong R.
    Jiqiren/Robot, 2016, 38 (04): : 451 - 457
  • [33] REINFORCEMENT LEARNING FOR ENERGY-EFFICIENT WIRELESS TRANSMISSION
    Mastronarde, Nicholas
    van der Schaar, Mihaela
    2011 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2011, : 3452 - 3455
  • [34] Parallel Deep Reinforcement Learning Method for Gait Control of Biped Robot
    Tao, Chongben
    Xue, Jie
    Zhang, Zufeng
    Gao, Zhen
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2022, 69 (06) : 2802 - 2806
  • [35] Natural policy gradient reinforcement learning for a CPG control of a biped robot
    Nakamura, Y
    Mori, T
    Ishii, S
    PARALLEL PROBLEM SOLVING FROM NATURE - PPSN VIII, 2004, 3242 : 972 - 981
  • [36] LEARNING TO CONTROL THE THREE-LINK MUSCULOSKELETAL ARM USING ACTOR-CRITIC REINFORCEMENT LEARNING ALGORITHM DURING REACHING MOVEMENT
    Tahami, Ehsan
    Jafari, Amir Homayoun
    Fallah, Ali
    BIOMEDICAL ENGINEERING-APPLICATIONS BASIS COMMUNICATIONS, 2014, 26 (05):
  • [37] DeepComfort: Energy-Efficient Thermal Comfort Control in Buildings Via Reinforcement Learning
    Gao, Guanyu
    Li, Jie
    Wen, Yonggang
    IEEE INTERNET OF THINGS JOURNAL, 2020, 7 (09) : 8472 - 8484
  • [38] An adaptive energy-efficient link layer protocol using stochastic learning control
    Kiran, S
    Chandramouli, R
    2003 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS, VOLS 1-5: NEW FRONTIERS IN TELECOMMUNICATIONS, 2003, : 1114 - 1118
  • [39] LEARNING ROBUST AND ENERGY-EFFICIENT BIPED WALKING PATTERNS USING QWALKING
    Ma, J.
    Cameron, S. A.
    MOBILE ROBOTICS-SOLUTIONS AND CHALLENGES, 2010, : 599 - 606
  • [40] Learning an Efficient Gait Cycle of a Biped Robot Based on Reinforcement Learning and Artificial Neural Networks
    Gil, Cristyan R.
    Calvo, Hiram
    Sossa, Humberto
    APPLIED SCIENCES-BASEL, 2019, 9 (03):