Neural Q-Learning Based on Residual Gradient for Nonlinear Control Systems

被引:0
|
作者
Si, Yanna [1 ]
Pu, Jiexin [1 ]
Zang, Shaofei [1 ]
机构
[1] Henan Univ Sci & Technol, Sch Informat Engn, Luoyang, Peoples R China
关键词
Q-learning; feedforward neural network; value function approximation; residual gradient method; nonlinear control systems;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
To solve the control problem of nonlinear system under continuous state space, this paper puts forward a neural Q-learning algorithm based on residual gradient method. Firstly, the multi-layer feedforward neural network is utilized to approximate the Q-value function, overcoming the "dimensional disaster" in the classical reinforcement learning. Then based on the residual gradient method, a mini-batch gradient descent is implemented by the experience replay to update the neural network parameters, which can effectively reduce the iterations number and increase the learning speed. Moreover, the momentum optimization method is introduced to ensure the stability of the training process further and improve the convergence. In order to balance exploration and utilization better, epsilon-decreasing strategy replaces epsilon-greedy for action selection. The simulation results of CartPole control task show the correctness and effectiveness of the proposed algorithm.
引用
收藏
页数:5
相关论文
共 50 条
  • [1] Elevator group control algorithm based on residual gradient and Q-learning
    Zong, ZL
    Wang, XG
    Tang, Z
    Zeng, GZ
    SICE 2004 ANNUAL CONFERENCE, VOLS 1-3, 2004, : 329 - 331
  • [2] Q-Learning Based Parameter Tuning for Model-free Adaptive Control of Nonlinear Systems
    Xu, Liuyong
    Hao, Shoulin
    Liu, Tao
    Zhu, Yong
    Wang, Haixia
    Zhang, Jiyan
    2024 14TH ASIAN CONTROL CONFERENCE, ASCC 2024, 2024, : 2078 - 2083
  • [3] Neural Q-learning
    Stephan ten Hagen
    Ben Kröse
    Neural Computing & Applications, 2003, 12 : 81 - 88
  • [4] Neural Q-learning
    ten Hagen, S
    Kröse, B
    NEURAL COMPUTING & APPLICATIONS, 2003, 12 (02): : 81 - 88
  • [5] Fuzzy neural control of systems with unknown dynamic using Q-learning strategies
    Kwok, DP
    Deng, ZD
    Li, CK
    Leung, TP
    Sun, ZQ
    Wong, JCK
    PROCEEDINGS OF THE 12TH IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS, VOLS 1 AND 2, 2003, : 482 - 487
  • [6] Neural-network-based accelerated safe Q-learning for optimal control of discrete-time nonlinear systems with state constraints☆
    Zhao, Mingming
    Wang, Ding
    Qiao, Junfei
    NEURAL NETWORKS, 2025, 186
  • [7] Fuzzy Q-learning Control for Temperature Systems
    Chen, Yeong-Chin
    Hung, Lon-Chen
    Syamsudin, Mariana
    22ND IEEE/ACIS INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING, ARTIFICIAL INTELLIGENCE, NETWORKING AND PARALLEL/DISTRIBUTED COMPUTING (SNPD 2021-FALL), 2021, : 148 - 151
  • [8] Fidelity-Based Probabilistic Q-Learning for Control of Quantum Systems
    Chen, Chunlin
    Dong, Daoyi
    Li, Han-Xiong
    Chu, Jian
    Tarn, Tzyh-Jong
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2014, 25 (05) : 920 - 933
  • [9] On-line fuzzy neural control of satellite attitude based on Q-learning
    College of Astronautics, Northwestern Polytechnical University, Xi'an 710072, China
    不详
    不详
    Beijing Ligong Daxue Xuebao, 2006, 3 (226-229):
  • [10] Optimal Tracking Control of Nonlinear Multiagent Systems Using Internal Reinforce Q-Learning
    Peng, Zhinan
    Luo, Rui
    Hu, Jiangping
    Shi, Kaibo
    Nguang, Sing Kiong
    Ghosh, Bijoy Kumar
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (08) : 4043 - 4055