Offline Reinforcement Learning of Robotic Control Using Deep Kinematics and Dynamics

被引:3
|
作者
Li, Xiang [1 ]
Shang, Weiwei [1 ]
Cong, Shuang [1 ]
机构
[1] Univ Sci & Technol China, Dept Automat, Hefei 230027, Peoples R China
基金
中国国家自然科学基金;
关键词
Computed-torque controller; kinematic and dynamic model learning; model-based reinforcement learning (MBRL); robotic control; trajectory tracking; NEURAL-NETWORKS;
D O I
10.1109/TMECH.2023.3336316
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
With the rapid development of deep learning, model-free reinforcement learning algorithms have achieved remarkable results in many fields. However, their high sample complexity and the potential for causing damage to environments and robots pose severe challenges for their application in real-world environments. Model-based reinforcement learning algorithms are often used to reduce the sample complexity. One limitation of these algorithms is the inevitable modeling errors. While the black-box model can fit complex state transition models, it ignores the existing knowledge of physics and robotics, especially studies of kinematic and dynamic models of the robotic manipulator. Compared with the black-box model, the physics-inspired deep models do not require specific knowledge of each system to obtain interpretable kinematic and dynamic models. In model-based reinforcement learning, these models can simulate the motion and be combined with classical controllers. This is due to their sharing the same form as traditional models, leading to higher precision tracking results. In this work, we utilize physics-inspired deep models to learn the kinematics and dynamics of a robotic manipulator. We propose a model-based offline reinforcement learning algorithm for controller parameter learning, combined with the traditional computed-torque controller. Experiments on trajectory tracking control of the Baxter manipulator, both in joint and operational space, are conducted in simulation and real environments. Experimental results demonstrate that our algorithm can significantly improve tracking accuracy and exhibits strong generalization and robustness.
引用
收藏
页码:2428 / 2439
页数:12
相关论文
共 50 条
  • [41] Mild Hybrid Electric Vehicle Powertrain Control Using Offline-Online Hybrid Deep Reinforcement Learning Strategy
    Yao, Zhengyu
    Yoon, Hwan-Sik
    SAE INTERNATIONAL JOURNAL OF ELECTRIFIED VEHICLES, 2023, 12 (03): : 331 - 341
  • [42] Nonlinear Optimal Control Using Deep Reinforcement Learning
    Bucci, Michele Alessandro
    Semeraro, Onofrio
    Allauzen, Alexandre
    Cordier, Laurent
    Mathelin, Lionel
    IUTAM LAMINAR-TURBULENT TRANSITION, 2022, 38 : 279 - 290
  • [43] CNC machine control using deep reinforcement learning
    Kalandyk, Dawid
    Kwiatkowski, Bogdan
    Mazur, Damian
    BULLETIN OF THE POLISH ACADEMY OF SCIENCES-TECHNICAL SCIENCES, 2024, 72 (03)
  • [44] Quadrotor motion control using deep reinforcement learning
    Jiang, Zifei
    Lynch, Alan F.
    JOURNAL OF UNMANNED VEHICLE SYSTEMS, 2021, 9 (04) : 234 - 251
  • [45] Manipulator Control using Federated Deep Reinforcement Learning
    Shivkumar, S.
    Kumaar, A. A. Nippun
    10TH INTERNATIONAL CONFERENCE ON ELECTRONICS, COMPUTING AND COMMUNICATION TECHNOLOGIES, CONECCT 2024, 2024,
  • [46] Offline Deep Reinforcement Learning for Dynamic Pricing of Consumer Credit
    Khraishi, Raad
    Okhrati, Ramin
    3RD ACM INTERNATIONAL CONFERENCE ON AI IN FINANCE, ICAIF 2022, 2022, : 325 - 333
  • [47] Dynamic metasurface control using Deep Reinforcement Learning
    Zhao, Ying
    Li, Liang
    Lanteri, Stephane
    Viquerat, Jonathan
    MATHEMATICS AND COMPUTERS IN SIMULATION, 2022, 197 : 377 - 395
  • [48] Overcoming model bias for robust offline deep reinforcement learning
    Swazinna, Phillip
    Udluft, Steffen
    Runkler, Thomas
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2021, 104
  • [49] Comparison of online and offline deep reinforcement learning with model predictive control for thermal energy management
    Brandi, Silvio
    Fiorentini, Massimo
    Capozzoli, Alfonso
    AUTOMATION IN CONSTRUCTION, 2022, 135
  • [50] Learning inverse kinematics and dynamics of a robotic manipulator using generative adversarial networks
    Ren, Hailin
    Ben-Tzvi, Pinhas
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2020, 124