TEFNet: Target-Aware Enhanced Fusion Network for RGB-T Tracking

被引:0
|
作者
Chen, Panfeng [1 ]
Gong, Shengrong [2 ]
Ying, Wenhao [2 ]
Du, Xin [3 ]
Zhong, Shan [2 ]
机构
[1] Huzhou Univ, Sch Informat Engn, Huzhou 313000, Peoples R China
[2] Changshu Inst Technol, Sch Comp Sci & Engn, Suzhou 215500, Peoples R China
[3] Suzhou Univ Sci & Technol, Sch Elect & Informat Engn, Suzhou 215009, Peoples R China
基金
中国博士后科学基金; 中国国家自然科学基金;
关键词
RGB-T tracking; Background elimination; Complementary information;
D O I
10.1007/978-981-99-8549-4_36
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
RGB-T tracking leverages the fusion of visible (RGB) and thermal (T) modalities to achieve more robust object tracking. Existing popular RGBT trackers often fail to fully leverage background information and complementary information from different modalities. To address these issues, we propose the target-aware enhanced fusion network (TEFNet). TEFNet concatenates the features of template and search regions from each modality and then utilizes self-attention operations to enhance the single-modality features for the target by discriminating it from the background. Additionally, a background elimination module is introduced to reduce the background regions. To further fuse the complementary information across different modalities, a dual-layer fusion module based on channel attention, self-attention, and bidirectional cross-attention is constructed. This module diminishes the feature information of the inferior modality, and amplifies the feature information of the dominant modality, effectively eliminating the adverse effects caused by modality differences. Experimental results on the LasHeR and VTUAV datasets demonstrate that our method outperforms other representative RGB-T tracking approaches, with significant improvements of 6.6% and 7.1% in MPR and MSR on the VTUAV dataset respectively.
引用
收藏
页码:432 / 443
页数:12
相关论文
共 50 条
  • [41] Anchor free based Siamese network tracker with transformer for RGB-T tracking
    Fan, Liangsong
    Kim, Pyeoungkee
    SCIENTIFIC REPORTS, 2023, 13 (01)
  • [42] Anchor free based Siamese network tracker with transformer for RGB-T tracking
    Liangsong Fan
    Pyeoungkee Kim
    Scientific Reports, 13
  • [43] Weighted Guided Optional Fusion Network for RGB-T Salient Object Detection
    Wang, Jie
    Li, Guoqiang
    Shi, Jie
    Xi, Jinwen
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2024, 20 (05)
  • [44] Learning Multi-domain Convolutional Network for RGB-T Visual Tracking
    Zhang, Xingming
    Zhang, Xuehan
    Du, Xuedan
    Zhou, Xiangming
    Yin, Jun
    2018 11TH INTERNATIONAL CONGRESS ON IMAGE AND SIGNAL PROCESSING, BIOMEDICAL ENGINEERING AND INFORMATICS (CISP-BMEI 2018), 2018,
  • [45] Unsupervised RGB-T object tracking with attentional multi-modal feature fusion
    Shenglan Li
    Rui Yao
    Yong Zhou
    Hancheng Zhu
    Bing Liu
    Jiaqi Zhao
    Zhiwen Shao
    Multimedia Tools and Applications, 2023, 82 : 23595 - 23613
  • [46] DHFNet: Decoupled Hierarchical Fusion Network for RGB-T dense prediction tasks
    Chen, Haojie
    Wang, Zhuo
    Qin, Hongde
    Mu, Xiaokai
    NEUROCOMPUTING, 2024, 583
  • [47] Cross-Collaboration Weighted Fusion Network for RGB-T Salient Detection
    Wang, Yumei
    Dongye, Changlei
    Zhao, Wenxiu
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT IV, ICIC 2024, 2024, 14865 : 301 - 312
  • [48] DaCFN: divide-and-conquer fusion network for RGB-T object detection
    Bofan Wang
    Haitao Zhao
    Yi Zhuang
    International Journal of Machine Learning and Cybernetics, 2023, 14 : 2407 - 2420
  • [49] Light-sensitive and adaptive fusion network for RGB-T crowd counting
    Huang, Liangjun
    Kang, Wencan
    Chen, Guangkai
    Zhang, Qing
    Zhang, Jianwei
    VISUAL COMPUTER, 2024, 40 (10): : 7279 - 7292
  • [50] MFCNet: Multimodal Feature Fusion Network for RGB-T Vehicle Density Estimation
    Qin, Ling-Xiao
    Sun, Hong-Mei
    Duan, Xiao-Meng
    Che, Cheng-Yue
    Jia, Rui-Sheng
    IEEE INTERNET OF THINGS JOURNAL, 2025, 12 (04): : 4207 - 4219