UAPT: an underwater acoustic target recognition method based on pre-trained Transformer

被引:0
|
作者
Tang, Jun [1 ]
Ma, Enxue [1 ]
Qu, Yang [1 ]
Gao, Wenbo [1 ]
Zhang, Yuchen [1 ]
Gan, Lin [2 ]
机构
[1] Tianjin Univ, Sch Civil Engn, Tianjin 300072, Peoples R China
[2] Northwestern Polytech Univ, Sch Automat, Xian 710072, Peoples R China
关键词
Underwater acoustic target recognition; Transformer; Transfer learning; Deep learning; Pre-train; MODEL;
D O I
10.1007/s00530-024-01614-3
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The Convolutional Neural Network (CNN) model in underwater acoustic target recognition (UATR) research reveals limitations arising from its inability to capture long-distance dependencies, impeding its capacity to focus on global information within the underwater acoustic signal. In contrast, the Transformer model has progressively emerged as the optimal choice in various studies, owing to its exclusive dependence on the attention mechanism for extracting global features from input data. Limited research utilizing the Transformer model in UATR has relied on an early ViT model, while in this paper, two refined Transformer models, namely Swin Transformer and Biformer, are adopted as the foundational networks, and a novel Swin Biformer model is proposed by harnessing the strengths of the two. Experimental results demonstrate the consistent superiority of the three models over CNN and ViT in UATR, and the Swin Biformer model remarkably attains the highest recognition accuracy of 94.3% evaluated on a dataset constructed from the Deepship database. At the same time, this paper proposes a UATR method based on pre-trained Transformer, the effectiveness of which is underscored by experimental findings as a recognition accuracy of approximately 97% was achieved on a generalized dataset derived from the Shipsear database. Even with limited data samples and more stringent classification requirements, the method maintains a recognition accuracy of over 90%, all while significantly reducing the training duration.
引用
收藏
页数:15
相关论文
共 50 条
  • [21] Integrally Pre-Trained Transformer Pyramid Networks
    Tian, Yunjie
    Xie, Lingxi
    Wang, Zhaozhi
    Wei, Longhui
    Zhang, Xiaopeng
    Jiao, Jianbin
    Wang, Yaowei
    Tian, Qi
    Ye, Qixiang
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 18610 - 18620
  • [22] Simple and Effective Multimodal Learning Based on Pre-Trained Transformer Models
    Miyazawa, Kazuki
    Kyuragi, Yuta
    Nagai, Takayuki
    IEEE ACCESS, 2022, 10 : 29821 - 29833
  • [23] Pre-trained Transformer-based Classification for Automated Patentability Examination
    Lo, Hao-Cheng
    Chu, Jung-Mei
    2021 IEEE ASIA-PACIFIC CONFERENCE ON COMPUTER SCIENCE AND DATA ENGINEERING (CSDE), 2021,
  • [24] A Novel Underwater Acoustic Target Recognition Method Based on MFCC and RACNN
    Liu, Dali
    Yang, Hongyuan
    Hou, Weimin
    Wang, Baozhu
    SENSORS, 2024, 24 (01)
  • [25] An Underwater Acoustic Target Recognition Method Based on Restricted Boltzmann Machine
    Luo, Xinwei
    Feng, Yulin
    SENSORS, 2020, 20 (18) : 1 - 18
  • [26] Underwater acoustic target recognition method based on a joint neural network
    Han, Xing Cheng
    Ren, Chenxi
    Wang, Liming
    Bai, Yunjiao
    PLOS ONE, 2022, 17 (04):
  • [27] An Underwater Acoustic Target Recognition Method Based on Spectrograms with Different Resolutions
    Luo, Xinwei
    Zhang, Minghong
    Liu, Ting
    Huang, Ming
    Xu, Xiaogang
    JOURNAL OF MARINE SCIENCE AND ENGINEERING, 2021, 9 (11)
  • [28] Biomedical generative pre-trained based transformer language model for age-related disease target discovery
    Zagirova, Diana
    Pushkov, Stefan
    Leung, Geoffrey Ho Duen
    Liu, Bonnie Hei Man
    Urban, Anatoly
    Sidorenko, Denis
    Kalashnikov, Aleksandr
    Kozlova, Ekaterina
    Naumov, Vladimir
    Pun, Frank W.
    Ozerov, Ivan V.
    Aliper, Alex
    Zhavoronkov, Alex
    AGING-US, 2023, 15 (18): : 9293 - 9309
  • [29] Detecting Syntactic Change with Pre-trained Transformer Models
    Hou, Liwen
    Smith, David A.
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 3564 - 3574
  • [30] Generative Pre-Trained Transformer for Cardiac Abnormality Detection
    Gaudilliere, Pierre Louis
    Sigurthorsdottir, Halla
    Aguet, Clementine
    Van Zaen, Jerome
    Lemay, Mathieu
    Delgado-Gonzalo, Ricard
    2021 COMPUTING IN CARDIOLOGY (CINC), 2021,