Reinforcement Learning-based path tracking for underactuated UUV under intermittent communication

被引:4
|
作者
Liu Z. [1 ]
Cai W. [1 ]
Zhang M. [2 ]
机构
[1] Hangzhou Dianzi University, 2nd Street, Zhejiang, Hangzhou
[2] Zhejiang University of Water Resources and Electric Power, 2nd Street, Zhejiang, Hangzhou
基金
中国国家自然科学基金;
关键词
Intermittent communication; Path control; Self-attention mechanism; Soft Actor and Critic (SAC); Unmanned Underwater Vehicle (UUV);
D O I
10.1016/j.oceaneng.2023.116076
中图分类号
学科分类号
摘要
This paper studies the path control of a six-degree-of-freedom underactuated Unmanned Underwater Vehicle (UUV) under limited communication conditions. Considering the large number of coupling between six-degree-of-freedom underactuated UUV of unknown dynamic models, traditional model-based control methods are difficult to effectively solve the three-dimensional path control problem. A self-attention based soft actor and critic (A-SAC) algorithm is designed to learn effective control policy from random paths. The problem of limited target acquisition by UUV in the actual underwater environment is effectively overcome, which is mainly caused by the inability of UUV to consistently receive information about their expected path. A new state space is designed and a self-attention mechanism is introduced to improve the efficiency of using discontinuous path information. Furthermore, the validation experiment compares classical Reinforcement Learning methods such as DDPG, PPO, and etc. Compared to other existing methods, the proposed A-SAC algorithm can more quickly and effectively learn the path control policy for a six-degree-of-freedom UUV that operates in a complex environment. © 2023 Elsevier Ltd
引用
收藏
相关论文
共 50 条
  • [21] A Heuristically Accelerated Reinforcement Learning-Based Neurosurgical Path Planner
    Ji, Guanglin
    Gao, Qian
    Zhang, Tianwei
    Cao, Lin
    Sun, Zhenglong
    CYBORG AND BIONIC SYSTEMS, 2023, 4
  • [22] A Reinforcement Learning-Based Path Planning Considering Degree of Observability
    Cho, Yong Hyeon
    Park, Chan Gook
    PROCEEDINGS OF THE 2020 INTERNATIONAL CONFERENCE ON ARTIFICIAL LIFE AND ROBOTICS (ICAROB2020), 2020, : 502 - 505
  • [23] A Heuristically Accelerated Reinforcement Learning-Based Neurosurgical Path Planner
    Ji G.
    Gao Q.
    Zhang T.
    Cao L.
    Sun Z.
    Cyborg and Bionic Systems, 2023, 4
  • [24] Smooth reinforcement learning-based trajectory tracking for articulated vehicles
    Chen, Liangfa
    Song, Xujie
    Xiao, Liming
    Gao, Lulu
    Zhang, Fawang
    Li, Shengbo
    Ma, Fei
    Duan, Jingliang
    Harbin Gongye Daxue Xuebao/Journal of Harbin Institute of Technology, 2024, 56 (12): : 116 - 123
  • [25] Asymptotic tracking by a reinforcement learning-based adaptive critic controller
    Shubhendu BHASIN
    Nitin SHARMA
    Parag PATRE
    Warren DIXON
    JournalofControlTheoryandApplications, 2011, 9 (03) : 400 - 409
  • [26] Asymptotic tracking by a reinforcement learning-based adaptive critic controller
    Bhasin S.
    Sharma N.
    Patre P.
    Dixon W.
    Journal of Control Theory and Applications, 2011, 9 (3): : 400 - 409
  • [27] Semiparametric Musculoskeletal Model for Reinforcement Learning-Based Trajectory Tracking
    Xu, Haoran
    Fan, Jianyin
    Ma, Hongxu
    Wang, Qiang
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2024, 73 : 1 - 16
  • [28] Reinforcement learning-based tracking control for AUVs subject to disturbances
    Wang, Guangcang
    Zhang, Dianfeng
    Wu, Zhaojing
    2022 34TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2022, : 2222 - 2227
  • [29] Reinforcement Learning-Based Tracking Control For Wheeled Mobile Robot
    Nguyen Tan Luy
    PROCEEDINGS 2012 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2012, : 462 - 467
  • [30] Reinforcement learning-based tracking control for a quadrotor unmanned aerial vehicle under external disturbances
    Liu, Hui
    Li, Bo
    Xiao, Bing
    Ran, Dechao
    Zhang, Chengxi
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2023, 33 (17) : 10360 - 10377