Sparse Transformer-Based Sequence Generation for Visual Object Tracking

被引:0
|
作者
Tian, Dan [1 ]
Liu, Dong-Xin [2 ]
Wang, Xiao [2 ]
Hao, Ying [2 ]
机构
[1] Shenyang Univ, Sch Intelligent Syst Sci & Engn, Shenyang 110044, Liaoning, Peoples R China
[2] Shenyang Univ, Sch Informat Engn, Shenyang 110044, Liaoning, Peoples R China
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Transformers; Visualization; Target tracking; Decoding; Feature extraction; Attention mechanisms; Object tracking; Training; Interference; Attention mechanism; sequence generation; sparse attention; visual object tracking; vision transformer;
D O I
10.1109/ACCESS.2024.3482468
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In visual object tracking, attention mechanisms can flexibly and efficiently handle complex dependencies and global information, which improves tracking accuracy. However, when dealing with scenarios that contain a large amount of background information or other complex information, its global attention ability can dilute the weight of important information, allocate unnecessary attention to background information, and thus reduce tracking performance. To relieve this problem, this paper proposes a visual object tracking framework based on a sparse transformer. Our tracking framework is a simple encoder-decoder structure that realizes the prediction of the target in an autoregressive manner, eliminating the additional head network and simplifying the tracking architecture. Furthermore, we introduce a Sparse Attention Mechanism (SMA) in the cross-attention layer of the decoder. Unlike traditional attention mechanisms, SMA focuses only on the top K pixel values that are most relevant to the current pixel when calculating attention weights. This allows the model to focus more on key information and improve foreground and background discrimination, resulting in more accurate and robust tracking. We conduct tests on six tracking benchmarks, and the experimental results prove the effectiveness of our method.
引用
收藏
页码:154418 / 154425
页数:8
相关论文
共 50 条
  • [31] Neural Rule-Execution Tracking Machine For Transformer-Based Text Generation
    Wang, Yufei
    Xu, Can
    Hu, Huang
    Tao, Chongyang
    Wan, Stephen
    Dras, Mark
    Johnson, Mark
    Jiang, Daxin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [32] Multiple templates transformer for visual object tracking
    Pang, Haibo
    Su, Jie
    Ma, Rongqi
    Li, Tingting
    Liu, Chengming
    KNOWLEDGE-BASED SYSTEMS, 2023, 280
  • [33] A Transformer-Based Framework for Tiny Object Detection
    Liao, Yi-Kai
    Lin, Gong-Si
    Yeh, Mei-Chen
    2023 ASIA PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE, APSIPA ASC, 2023, : 373 - 377
  • [34] Survey of Transformer-Based Object Detection Algorithms
    Li, Jian
    Du, Jianqiang
    Zhu, Yanchen
    Guo, Yongkun
    Computer Engineering and Applications, 2023, 59 (10) : 48 - 64
  • [35] TransGOP: Transformer-Based Gaze Object Prediction
    Wang, Binglu
    Guo, Chenxi
    Jin, Yang
    Xia, Haisheng
    Liu, Nian
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 9, 2024, : 10180 - 10188
  • [36] A GENERAL BAYESIAN ALGORITHM FOR VISUAL OBJECT TRACKING BASED ON SPARSE FEATURES
    Soto, Mauricio
    Regazzoni, Carlo S.
    2011 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2011, : 1181 - 1184
  • [37] TFMFT: Transformer-based multiple fish tracking
    Li, Weiran
    Liu, Yeqiang
    Wang, Wenxu
    Li, Zhenbo
    Yue, Jun
    COMPUTERS AND ELECTRONICS IN AGRICULTURE, 2024, 217
  • [38] Transformer-based convolutional forgetting knowledge tracking
    Tieyuan Liu
    Meng Zhang
    Chuangying Zhu
    Liang Chang
    Scientific Reports, 13
  • [39] Transformer-based convolutional forgetting knowledge tracking
    Liu, Tieyuan
    Zhang, Meng
    Zhu, Chuangying
    Chang, Liang
    SCIENTIFIC REPORTS, 2023, 13 (01)
  • [40] Transformer-Based Multiple-Object Tracking via Anchor-Based-Query and Template Matching
    Wang, Qinyu
    Lu, Chenxu
    Gao, Long
    He, Gang
    SENSORS, 2024, 24 (01)