Hierarchical residual reinforcement learning based path tracking control method for unmanned bicycle

被引:0
|
作者
Huo, Benyan [1 ]
Yu, Long [1 ]
Liu, Yanhong [1 ]
Chen, Zhang [2 ]
机构
[1] School of Electrical and Information Engineering, Zhengzhou University, Zhengzhou,450001, China
[2] Department of Automation, Tsinghua University, Beijing,100084, China
基金
中国国家自然科学基金;
关键词
Attitude control - Bicycles - Reinforcement learning;
D O I
10.1016/j.robot.2025.104996
中图分类号
学科分类号
摘要
Due to their super under-actuation and high nonlinearity properties, simplification and linearization techniques are necessary in the traditional model-based path tracking controller design of unmanned bicycles, leading to the decrease of their control performance. On the other hand, as one of the most important learning-based methods, deep reinforcement learning (DRL) suffers from low training efficiency and initial safety issues. In this letter, a hierarchical residual reinforcement learning (HRRL)-based path tracking control method is proposed, to address the drawbacks of both traditional and learning-based approaches. The path tracking task is decomposed into two subtasks, i.e., attitude control and position control, and the controllers are designed separately for each subtask. In each controller, a DRL controller is connected to a traditional controller through residual connection. Physical simulation experiments demonstrate that compared to the traditional LQR, LQI, Stanley and DRL approaches, the proposed method can improve the tracking performance of unmanned bicycles and decrease the training time and the tipping number during training. Furthermore, experimental results also show that the proposed controller exhibits a certain level of robustness, enabling effective path tracking in complex terrain. © 2025 Elsevier B.V.
引用
收藏
相关论文
共 50 条
  • [41] A novel path planning approach for unmanned ships based on deep reinforcement learning
    Chen, Chen
    Ma, Feng
    Liu, Jia-Lun
    Yan, Xin-Ping
    Chen, Xian-Qiao
    DATA SCIENCE AND KNOWLEDGE ENGINEERING FOR SENSING DECISION SUPPORT, 2018, 11 : 626 - 633
  • [42] Hierarchical reinforcement learning using path clustering
    Gil, Paulo
    Nunes, Luis
    PROCEEDINGS OF THE 2013 8TH IBERIAN CONFERENCE ON INFORMATION SYSTEMS AND TECHNOLOGIES (CISTI 2013), 2013,
  • [43] Optimization Method of LQR Curve Path Tracking Controller for Unmanned Rice Transplanter Based on Fuzzy Control
    Ma Y.
    Chi R.
    Zhao Y.
    Ban C.
    Su T.
    Li Z.
    Nongye Jixie Xuebao/Transactions of the Chinese Society for Agricultural Machinery, 2023, 54 : 1 - 8and102
  • [44] Hierarchical Evasive Path Planning Using Reinforcement Learning and Model Predictive Control
    Feher, Arpad
    Aradi, Szilard
    Becsi, Tamas
    IEEE ACCESS, 2020, 8 : 187470 - 187482
  • [45] A Hierarchical Resource Scheduling Method for Satellite Control System Based on Deep Reinforcement Learning
    Li, Yang
    Guo, Xiye
    Meng, Zhijun
    Qin, Junxiang
    Li, Xuan
    Ma, Xiaotian
    Ren, Sichuang
    Yang, Jun
    ELECTRONICS, 2023, 12 (19)
  • [46] Reinforcement learning-based trajectory tracking optimal control of unmanned surface vehicles in narrow water areas
    Wei, Ziping
    Du, Jialu
    ISA TRANSACTIONS, 2025, 159 : 152 - 164
  • [47] Reinforcement learning of composite disturbance observer based tracking control for unmanned aerial helicopter under outside disturbances
    Zhang, Chunyu
    Lu, Changyu
    Li, Tao
    Mao, Zehui
    AEROSPACE SCIENCE AND TECHNOLOGY, 2025, 161
  • [48] Adaptive Optimal Tracking Control for Uncertain Unmanned Surface Vessel via Reinforcement Learning
    Chen, Lin
    Wang, Min
    Dai, Shi-Lu
    2021 PROCEEDINGS OF THE 40TH CHINESE CONTROL CONFERENCE (CCC), 2021, : 8398 - 8403
  • [49] Safe tracking control for unmanned aerial helicopter under input saturation based on reinforcement learning and disturbance observer
    Zhang, Chunyu
    Li, Tao
    Ye, Yongqiang
    Mao, Zehui
    NEUROCOMPUTING, 2025, 630
  • [50] A Target- Aware Well Path Control Method Based on Transfer Reinforcement Learning
    Zhu, Dandan
    Xu, Qiuhan
    Wang, Fei
    Chen, Dong
    Ye, Zhihui
    Zhou, Hao
    Zhang, Ke
    SPE JOURNAL, 2024, 29 (04): : 1730 - 1741