Automatic reinforcement for robust model-free neurocontrol of robots without persistent excitation

被引:0
|
作者
Pantoja-Garcia, Luis [1 ]
Parra-Vega, Vicente [1 ,3 ]
Garcia-Rodriguez, Rodolfo [2 ]
机构
[1] Ctr Res & Adv Studies, Robot & Adv Mfg, Saltillo, Mexico
[2] Univ Politecn Metropolitana Hidalgo, Aeronaut Engn Dept, Postgrad Program Aerosp Engn, Tolcayuca, Mexico
[3] Ave Ind Met 1062, Ramos Arizpe 25903, Mexico
关键词
automatic reinforced learning; model-free control; neurocontrol; persistent excitation; robot manipulators; TRACKING CONTROL; ADAPTIVE-CONTROL; MANIPULATOR CONTROL; NONLINEAR-SYSTEMS; NEURAL-NETWORKS; APPROXIMATION; FEEDBACK; CONVERGENCE;
D O I
10.1002/acs.3697
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Model-based adaptive control suffers over parametrization from the many adaptive parameters compared to the order of system dynamics, leading to sluggish tracking with a poor adaptation transient without robustness. Likewise, adaptive model-free neurocontrol that relies on the Stone-Weierstrass theorem also suffers from similar problems in addition to over-fitting to approximate inverse dynamics. This article proposes a novel reinforced adaptive mechanism to guarantee a transient and robustness for the model-free adaptive control of nonlinear Lagrangian systems. Inspired by the symbiosis of Actor-Critic (AC) architecture and integral sliding modes, the reinforced stage neural network, analogous to the critic, injects excitation signals to reinforce the parametric learning of the adaptive stage neural network, analogous to the actor to improve the approximation of inverse dynamics. The underlying integral sliding surface error drives improved learning onto a low-dimensional invariant manifold to guarantee local exponential convergence of tracking errors. Lyapunov stability substantiates the robustness with an improved transient response. Our proposal stands for a hybrid approach between AC and neurocontrol, where the reinforced stage does not require a value function nor reward to provide automatic reinforcement to the adaptive stage parametric adaptation. Dynamic simulations are presented for a nonlinear robot manipulator under different conditions.
引用
收藏
页码:221 / 236
页数:16
相关论文
共 50 条
  • [41] Model-based and Model-free Reinforcement Learning for Visual Servoing
    Farahmand, Amir Massoud
    Shademan, Azad
    Jagersand, Martin
    Szepesvari, Csaba
    ICRA: 2009 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, VOLS 1-7, 2009, : 4135 - 4142
  • [42] Model-Free Robust Backstepping Adaptive Cruise Control
    Zhang, Yanan
    Song, Jiacheng
    INTERNATIONAL JOURNAL OF AEROSPACE ENGINEERING, 2023, 2023
  • [43] A Model-Free Robust Control Approach for Robot Manipulator
    Izadbakhsh, A.
    Fateh, M. M.
    PROCEEDINGS OF WORLD ACADEMY OF SCIENCE, ENGINEERING AND TECHNOLOGY, VOL 23, 2007, 23 : 205 - 210
  • [44] Robust Model-Free Control Applied to a Quadrotor UAV
    Al Younes, Younes
    Drak, Ahmad
    Noura, Hassan
    Rabhi, Abdelhamid
    El Hajjaji, Ahmed
    JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 2016, 84 (1-4) : 37 - 52
  • [45] Robust Model-Free Control Applied to a Quadrotor UAV
    Younes Al Younes
    Ahmad Drak
    Hassan Noura
    Abdelhamid Rabhi
    Ahmed El Hajjaji
    Journal of Intelligent & Robotic Systems, 2016, 84 : 37 - 52
  • [46] Linear Quadratic Control Using Model-Free Reinforcement Learning
    Yaghmaie, Farnaz Adib
    Gustafsson, Fredrik
    Ljung, Lennart
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2023, 68 (02) : 737 - 752
  • [47] On Distributed Model-Free Reinforcement Learning Control With Stability Guarantee
    Mukherjee, Sayak
    Vu, Thanh Long
    IEEE CONTROL SYSTEMS LETTERS, 2021, 5 (05): : 1615 - 1620
  • [48] Model-Free Reinforcement Learning of Impedance Control in Stochastic Environments
    Stulp, Freek
    Buchli, Jonas
    Ellmer, Alice
    Mistry, Michael
    Theodorou, Evangelos A.
    Schaal, Stefan
    IEEE TRANSACTIONS ON AUTONOMOUS MENTAL DEVELOPMENT, 2012, 4 (04) : 330 - 341
  • [49] Model-Free Recurrent Reinforcement Learning for AUV Horizontal Control
    Huo, Yujia
    Li, Yiping
    Feng, Xisheng
    3RD INTERNATIONAL CONFERENCE ON AUTOMATION, CONTROL AND ROBOTICS ENGINEERING (CACRE 2018), 2018, 428
  • [50] Limit Reachability for Model-Free Reinforcement Learning of ω-Regular Objectives
    Hahn, Ernst Moritz
    Perez, Mateo
    Schewe, Sven
    Somenzi, Fabio
    Trivedi, Ashutosh
    Wojtczak, Dominik
    PROCEEDINGS OF THE 5TH INTERNATIONAL WORKSHOP ON SYMBOLIC-NUMERIC METHODS FOR REASONING ABOUT CPS AND IOT (SNR 2019), 2019, : 16 - 18