Gait Balance and Acceleration of a Biped Robot Based on Q-Learning

被引:37
|
作者
Lin, Jin-Ling [1 ]
Hwang, Kao-Shing [2 ]
Jiang, Wei-Cheng [2 ]
Chen, Yu-Jen [3 ]
机构
[1] Shih Hsin Univ, Dept Informat Management, Taipei 116, Taiwan
[2] Natl Sun Yat Sen Univ, Dept Elect Engn, Kaohsiung 80424, Taiwan
[3] Natl Chung Cheng Univ, Dept Elect Engn, Chiayi 62102, Taiwan
来源
IEEE ACCESS | 2016年 / 4卷
关键词
Reinforcement learning; biped robot; continuous action space; zero moment point; ALGORITHM;
D O I
10.1109/ACCESS.2016.2570255
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper presents a method for the biped dynamic walking and balance control using reinforcement learning, which learns dynamic walking without a priori knowledge about the dynamic model. The learning architecture developed is aimed to solve complex control problems in robotic actuation control by mapping the action space from a discretized domain to a continuous one. It employs the discrete actions to construct a policy for continuous action. The architecture allows for the scaling of the dimensionality of the state space and cardinality of the action set that represents new knowledge, or new requirements for a desired task. The balance learning method utilizing the motion of robot arm and leg to shift the zero moment point on the soles of a robot can maintain the biped robot in a static stable state. This balanced algorithm is applied to biped walking on a flat surface and a seesaw and is making the biped's walks more stable. The simulation shows that the proposed method can allow the robot to learn to improve its behavior in terms of walking speed. Finally, the methods are implemented on a physical biped robot to demonstrate the feasibility and effectiveness of the proposed learning scheme.
引用
收藏
页码:2439 / 2449
页数:11
相关论文
共 50 条
  • [41] ZMP-based Gait Optimization of the Biped Robot
    窦瑞军
    马培荪
    谢玲
    Journal of DongHua University, 2003, (04) : 83 - 86
  • [42] ZMP-based gait optimization of the biped robot
    Dou, Rui-Jun
    Ma, Pei-Sun
    Xie, Ling
    Journal of Dong Hua University (English Edition), 2003, 20 (04): : 83 - 86
  • [43] Optimization based gait pattern generation for a biped robot
    Buschmann, T
    Lohmeier, S
    Ulbrich, H
    Pfeiffer, F
    2005 5TH IEEE-RAS INTERNATIONAL CONFERENCE ON HUMANOID ROBOTS, 2005, : 98 - 103
  • [44] Q-Learning Based Selection Strategies for Load Balance and Energy Balance in Heterogeneous Networks
    Chen Jialing
    Yin Mingxi
    Duan Xiaohui
    Jiao Bingli
    2020 5TH INTERNATIONAL CONFERENCE ON COMPUTER AND COMMUNICATION SYSTEMS (ICCCS 2020), 2020, : 728 - 732
  • [45] Neural Q-Learning Controller for Mobile Robot
    Ganapathy, Velappa
    Yun, Soh Chin
    Joe, Halim Kusama
    2009 IEEE/ASME INTERNATIONAL CONFERENCE ON ADVANCED INTELLIGENT MECHATRONICS, VOLS 1-3, 2009, : 863 - 868
  • [46] Mobile Robot Navigation: Neural Q-Learning
    Yun, Soh Chin
    Parasuraman, S.
    Ganapathy, V.
    ADVANCES IN COMPUTING AND INFORMATION TECHNOLOGY, VOL 3, 2013, 178 : 259 - +
  • [47] Robot behavioral selection using Q-learning
    Martinson, E
    Stoytchev, A
    Arkin, R
    2002 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, VOLS 1-3, PROCEEDINGS, 2002, : 970 - 977
  • [48] Mobile robot navigation: neural Q-learning
    Parasuraman, S.
    Yun, Soh Chin
    INTERNATIONAL JOURNAL OF COMPUTER APPLICATIONS IN TECHNOLOGY, 2012, 44 (04) : 303 - 311
  • [49] Path Navigation For Indoor Robot With Q-Learning
    Huang, Lvwen
    He, Dongjian
    Zhang, Zhiyong
    Zhang, Peng
    INTELLIGENT AUTOMATION AND SOFT COMPUTING, 2016, 22 (02): : 317 - 323
  • [50] Autonomous Navigation based on a Q-learning algorithm for a Robot in a Real Environment
    Strauss, Clement
    Sahin, Ferat
    2008 IEEE INTERNATIONAL CONFERENCE ON SYSTEM OF SYSTEMS ENGINEERING (SOSE), 2008, : 361 - 365