Preference-Based Assistance Map Learning With Robust Adaptive Oscillators

被引:1
|
作者
Li, Shilei [1 ,2 ]
Zou, Wulin [1 ,2 ]
Duan, Pu [2 ]
Shi, Ling [1 ]
机构
[1] Hong Kong Univ Sci & Technol, Dept Elect & Comp Engn, Hong Kong, Peoples R China
[2] Xeno Dynam Co Ltd, Control Dept, Shenzhen 518055, Peoples R China
来源
关键词
Robust adaptive oscillators; Gaussian process regression; muscle activities; hip exoskeleton; HUMAN WALKING; GAIT; OPTIMIZATION; COST;
D O I
10.1109/TMRB.2022.3206609
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
Recently, lower-limb exoskeletons have demonstrated the ability to enhance human mobility by reducing biological efforts with human-in-the-loop (HIL) optimization. However, this technology is confined to the laboratory, and it is difficult to generalize to daily applications where gaits are more complex and professional equipment is not accessible. To solve this issue, firstly, we present a robust adaptive oscillator (RAO) to synchronize the human-robot movement and extract gait features. Then, we use the Gaussian process regression (GPR) to map the subjects' preferred assistance parameters to gait features. Experiments show that the RAO has a faster convergence rate compared with the traditional adaptive oscillators. Meanwhile, the learning efficiency of the proposed method shows superiority compared with the HIL optimization. The effectiveness of the proposed method is validated by a hip exoskeleton at a speed of 5 km/h with 7 participants. Three muscles which include rectus femoris, tibialis anterior, and medial gastrocnemius are investigated in three conditions: user-preferred assistance (ASS), zero torque (ZT), and normal walking (NW). The results show that all muscles achieve an activity reduction in ASS mode compared with ZT or NW. Meanwhile, there is a statistically significant difference on medial gastrocnemius in ASS mode with respect to both ZT and NW (-15.63 +/- 6.51 % and -8.73 +/- 6.40%, respectively).
引用
收藏
页码:1000 / 1009
页数:10
相关论文
共 50 条
  • [1] Preference-based learning to rank
    Nir Ailon
    Mehryar Mohri
    Machine Learning, 2010, 80 : 189 - 211
  • [2] Preference-Based Policy Learning
    Akrour, Riad
    Schoenauer, Marc
    Sebag, Michele
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, PT I, 2011, 6911 : 12 - 27
  • [3] Preference-based learning to rank
    Ailon, Nir
    Mohri, Mehryar
    MACHINE LEARNING, 2010, 80 (2-3) : 189 - 211
  • [4] Preference-based search with adaptive recommendations
    Viappiani, Paolo
    Pu, Pearl
    Faltings, Boi
    AI COMMUNICATIONS, 2008, 21 (2-3) : 155 - 175
  • [5] Preference-Based Policy Iteration: Leveraging Preference Learning for Reinforcement Learning
    Cheng, Weiwei
    Fuernkranz, Johannes
    Huellermeier, Eyke
    Park, Sang-Hyeun
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, PT I, 2011, 6911 : 312 - 327
  • [6] Learning state importance for preference-based reinforcement learning
    Zhang, Guoxi
    Kashima, Hisashi
    MACHINE LEARNING, 2023, 113 (4) : 1885 - 1901
  • [7] Learning state importance for preference-based reinforcement learning
    Guoxi Zhang
    Hisashi Kashima
    Machine Learning, 2024, 113 : 1885 - 1901
  • [8] Mixing corrupted preferences for robust and feedback-efficient preference-based reinforcement learning
    Heo, Jongkook
    Lee, Young Jae
    Kim, Jaehoon
    Kwak, Min Gu
    Park, Young Joon
    Kim, Seoung Bum
    KNOWLEDGE-BASED SYSTEMS, 2025, 309
  • [9] Preference-based reinforcement learning: evolutionary direct policy search using a preference-based racing algorithm
    Róbert Busa-Fekete
    Balázs Szörényi
    Paul Weng
    Weiwei Cheng
    Eyke Hüllermeier
    Machine Learning, 2014, 97 : 327 - 351
  • [10] Preference-based reinforcement learning: evolutionary direct policy search using a preference-based racing algorithm
    Busa-Fekete, Robert
    Szoerenyi, Balazs
    Weng, Paul
    Cheng, Weiwei
    Huellermeier, Eyke
    MACHINE LEARNING, 2014, 97 (03) : 327 - 351