Reinforcement Learning-based path tracking for underactuated UUV under intermittent communication

被引:4
|
作者
Liu Z. [1 ]
Cai W. [1 ]
Zhang M. [2 ]
机构
[1] Hangzhou Dianzi University, 2nd Street, Zhejiang, Hangzhou
[2] Zhejiang University of Water Resources and Electric Power, 2nd Street, Zhejiang, Hangzhou
基金
中国国家自然科学基金;
关键词
Intermittent communication; Path control; Self-attention mechanism; Soft Actor and Critic (SAC); Unmanned Underwater Vehicle (UUV);
D O I
10.1016/j.oceaneng.2023.116076
中图分类号
学科分类号
摘要
This paper studies the path control of a six-degree-of-freedom underactuated Unmanned Underwater Vehicle (UUV) under limited communication conditions. Considering the large number of coupling between six-degree-of-freedom underactuated UUV of unknown dynamic models, traditional model-based control methods are difficult to effectively solve the three-dimensional path control problem. A self-attention based soft actor and critic (A-SAC) algorithm is designed to learn effective control policy from random paths. The problem of limited target acquisition by UUV in the actual underwater environment is effectively overcome, which is mainly caused by the inability of UUV to consistently receive information about their expected path. A new state space is designed and a self-attention mechanism is introduced to improve the efficiency of using discontinuous path information. Furthermore, the validation experiment compares classical Reinforcement Learning methods such as DDPG, PPO, and etc. Compared to other existing methods, the proposed A-SAC algorithm can more quickly and effectively learn the path control policy for a six-degree-of-freedom UUV that operates in a complex environment. © 2023 Elsevier Ltd
引用
收藏
相关论文
共 50 条
  • [41] Bound Inference and Reinforcement Learning-based Path Construction in Bandwidth Tomography
    Feng, Cuiying
    An, Jianwei
    Wu, Kui
    Wang, Jianping
    IEEE CONFERENCE ON COMPUTER COMMUNICATIONS (IEEE INFOCOM 2021), 2021,
  • [42] Reinforcement learning-based dynamic obstacle avoidance and integration of path planning
    Choi, Jaewan
    Lee, Geonhee
    Lee, Chibum
    INTELLIGENT SERVICE ROBOTICS, 2021, 14 (05) : 663 - 677
  • [43] Bound Inference and Reinforcement Learning-Based Path Construction in Bandwidth Tomography
    Feng, Cuiying
    An, Jianwei
    Wu, Kui
    Wang, Jianping
    IEEE-ACM TRANSACTIONS ON NETWORKING, 2022, 30 (02) : 501 - 514
  • [44] Reinforcement learning-based dynamic obstacle avoidance and integration of path planning
    Jaewan Choi
    Geonhee Lee
    Chibum Lee
    Intelligent Service Robotics, 2021, 14 : 663 - 677
  • [45] Path-tracking control of underactuated ships under tracking error constraints
    Do K.D.
    Journal of Marine Science and Application, 2015, 14 (4) : 343 - 354
  • [46] Reinforcement Learning-Based Robust Tracking Control Application to Morphing Aircraft
    Yang, Zhicheng
    Tan, Junbo
    Wang, Xueqian
    Yao, Zongxin
    Liang, Bin
    2023 AMERICAN CONTROL CONFERENCE, ACC, 2023, : 2757 - 2762
  • [47] Reinforcement Learning-Based Data Association for Multiple Target Tracking in Clutter
    Qu, Chengzhi
    Zhang, Yan
    Zhang, Xin
    Yang, Yang
    SENSORS, 2020, 20 (22) : 1 - 29
  • [48] Reinforcement Learning-Based Underwater Acoustic Channel Tracking with Forgetting Factors
    Wang, Yuhang
    Li, Wei
    Hao, Zhonghan
    2022 OCEANS HAMPTON ROADS, 2022,
  • [49] Robust Reinforcement Learning-Based Tracking Control for Wheeled Mobile Robot
    Nguyen Tan Luy
    Nguyen Duc Thanh
    Nguyen Thien Thanh
    Nguyen Thi Phuong Ha
    2010 2ND INTERNATIONAL CONFERENCE ON COMPUTER AND AUTOMATION ENGINEERING (ICCAE 2010), VOL 1, 2010, : 171 - 176
  • [50] Reinforcement Learning-Based Tracking Control of USVs in Varying Operational Conditions
    Martinsen, Andreas B.
    Lekkas, Anastasios M.
    Gros, Sebastien
    Glomsrud, Jon Arne
    Pedersen, Tom Arne
    FRONTIERS IN ROBOTICS AND AI, 2020, 7