Deep Reinforcement Learning Based Dynamic Proportional-Integral (PI) Gain Auto-Tuning Method for a Robot Driver System

被引:6
|
作者
Park, Joonghoo [1 ]
Kim, Heejung [1 ]
Hwang, Kyunghun [2 ]
Lim, Sejoon [3 ]
机构
[1] Kookmin Univ, Grad Sch Automot Engn, Seoul 02707, South Korea
[2] Hyundai Motor Co, Electrificat Energy Efficiency & Drivabil Team 3, Hwaseong 18280, South Korea
[3] Kookmin Univ, Dept Automobile & IT Convergence, Seoul 02707, South Korea
来源
IEEE ACCESS | 2022年 / 10卷
基金
新加坡国家研究基金会;
关键词
Robots; Vehicles; PI control; Heuristic algorithms; Vehicle dynamics; Control systems; Dynamometers; Automation; deep Q-learning; emission test; machine learning; PID control; reinforcement learning; vehicle control; NEURAL-NETWORK CONTROL;
D O I
10.1109/ACCESS.2022.3159785
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
To meet the growing trend of stringent fuel economy regulations, automakers around the world are designing modules such as engines, motors, transmissions and batteries to be as efficient as possible. In order to verify the effect of these designs on the overall fuel efficiency of the vehicle, the vehicle equipped with each module is placed on the chassis dynamometer, driven to follow the target vehicle speed, and actual fuel efficiency is measured. These tests are traditionally performed by human operators, but are now being replaced by robots (physical or software) to ensure the accuracy and reliability of test results. Although the conventionally proposed proportional integral (PI)-based controller has a simple structure and is easy to implement, it requires the process of finding the optimal gain whenever the test conditions such as vehicle or drive cycle change, which is difficult and time consuming. In this study, we propose a proportional integral controller gain adjustment algorithm using deep reinforcement learning. The reinforcement learning agent learns to dynamically modify the PI gain value of the acceleration/deceleration pedal to better follow the target vehicle in a simulation environment. The perturbation is used in each training episode to reduce the difference between the simulation and real testing environment. Upon completion of the training process, the trained agent performs an adjustment process that generates a reference gain table. We then use this reference gain table to perform a real test. The performance of the proposed system was evaluated using Hyundai Tucson HEV (NX4) on an AVL chassis dynamometer. We also compared the performance of our proposed algorithm to traditional fuzzy logic-based PI controllers. The obtained experimental results show that the proposed control system achieved a performance improvement of aounrd 46.8% compared to the conventional PI control system in terms of root mean square error.
引用
收藏
页码:31043 / 31057
页数:15
相关论文
共 50 条
  • [31] Improved Robot Path Planning Method Based on Deep Reinforcement Learning
    Han, Huiyan
    Wang, Jiaqi
    Kuang, Liqun
    Han, Xie
    Xue, Hongxin
    SENSORS, 2023, 23 (12)
  • [32] Research on Robot Intelligent Control Method Based on Deep Reinforcement Learning
    Rao, Shu
    2022 6TH INTERNATIONAL SYMPOSIUM ON COMPUTER SCIENCE AND INTELLIGENT CONTROL, ISCSIC, 2022, : 221 - 225
  • [33] Navigation method for mobile robot based on hierarchical deep reinforcement learning
    Wang T.
    Li A.
    Song H.-L.
    Liu W.
    Wang M.-H.
    Kongzhi yu Juece/Control and Decision, 2022, 37 (11): : 2799 - 2807
  • [34] Deep reinforcement learning and decoupling proportional-integral-derivative control of a humanoid cable-driven hybrid robot
    Liu, Yong
    Luo, Zhisheng
    Qian, Sen
    Wang, Shuaikang
    Wu, Zhe
    INTERNATIONAL JOURNAL OF ADVANCED ROBOTIC SYSTEMS, 2024, 21 (03):
  • [35] Towards reliable robot packing system based on deep reinforcement learning
    Xiong, Heng
    Ding, Kai
    Ding, Wan
    Peng, Jian
    Xu, Jianfeng
    ADVANCED ENGINEERING INFORMATICS, 2023, 57
  • [36] Fuzzy auto-tuning scheme based on α-parameter ultimate sensitivity method for AC speed servo system
    Yoshitsugu, J
    Inoue, K
    Nakaoka, M
    CONFERENCE RECORD OF THE 1998 IEEE INDUSTRY APPLICATIONS CONFERENCE, VOLS 1-3, 1998, : 1625 - 1631
  • [37] Dynamic characteristics and deep reinforcement learning of proportional-integral-differential controller for quadruped stator-based ultrasonic linear motor
    Jiang, Yukun
    Wang, Fangyi
    Sasamura, Tatsuki
    Mustafa, Abdullah
    Morita, Takeshi
    JAPANESE JOURNAL OF APPLIED PHYSICS, 2024, 63 (04)
  • [38] Deep reinforcement learning-based proportional–integral control for dual-active-bridge converter
    Weiyu You
    Genke Yang
    Jian Chu
    Changjiang Ju
    Neural Computing and Applications, 2023, 35 : 17953 - 17966
  • [39] A Dynamic Proportional-Integral Observer-Based Nonlinear Fault-Tolerant Controller Design for Nonlinear System With Partially Unknown Dynamic
    Han, Jian
    Liu, Xiuhua
    Wei, Xinjiang
    Sun, Shaoxin
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2022, 52 (08): : 5092 - 5104
  • [40] Automated performance tuning of distributed storage system based on deep reinforcement learning
    Wang, Lu
    Zhang, Wentao
    Cheng, Yaodong
    19TH INTERNATIONAL WORKSHOP ON ADVANCED COMPUTING AND ANALYSIS TECHNIQUES IN PHYSICS RESEARCH, 2020, 1525