Multi-scale feature extraction and fusion with attention interaction for RGB-T

被引:1
|
作者
Xing, Haijiao [1 ]
Wei, Wei [1 ]
Zhang, Lei [1 ]
Zhang, Yanning [1 ]
机构
[1] Northwestern Polytech Univ, Sch Comp Sci, Xian 710072, Peoples R China
基金
中国国家自然科学基金;
关键词
Single-object tracking; RGB-T tracking; Feature fusion; SIAMESE NETWORKS; TRACKING;
D O I
10.1016/j.patcog.2024.110917
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
RGB-T single-object tracking aims to track objects utilizing both RGB images and thermal infrared(TIR) images. Though the siamese-based RGB-T tracker shows its advantage in tracking speed, its accuracy still cannot be compared with other state-of-the-art trackers (e.g., MDNet). In this study, we revisit the existing siamese-based RGB-T tracker and find that such fall behind comes from insufficient feature fusion between RGB image and TIR image, as well as incomplete interactions between template frame and search frame. Inspired by this, we propose a multi-scale feature extraction and fusion network with Temporal-Spatial Memory (MFATrack). Instead of fusing RGB image and TIR image with the single-scale feature map or only high-level features from the multi-scale feature map, MFATrack proposes a new fusion strategy by fusing features from all scales, which can capture contextual information in shallow layers and details in the deep layer. To learn the feature better for tracking tasks, MFATrack fuses the features via several consecutive frames. In addition, we also propose a self-attention interaction module specifically designed for the search frame, highlighting the features in the search frame that are relevant to the target and thus facilitating rapid convergence for target localization. Experimental results demonstrate the proposed MFATrack is not only fast, but also can obtain better tracking accuracy compared with other competing methods including MDNet-based methods and other siamese-based trackers.
引用
收藏
页数:12
相关论文
共 50 条
  • [21] Wavelet-Driven Multi-Band Feature Fusion for RGB-T Salient Object Detection
    Zhao, Jianxun
    Wen, Xin
    He, Yu
    Yang, Xiaowei
    Song, Kechen
    Sensors, 2024, 24 (24)
  • [22] L-Unet: A Landslide Extraction Model Using Multi-Scale Feature Fusion and Attention Mechanism
    Dong, Zhangyu
    An, Sen
    Zhang, Jin
    Yu, Jinqiu
    Li, Jinhui
    Xu, Daoli
    REMOTE SENSING, 2022, 14 (11)
  • [23] A multi-scale feature extraction and fusion method for bearing fault diagnosis based on hybrid attention mechanism
    Meng, Huan
    Zhang, Jiakai
    Zhao, Jingbo
    Wang, Daichao
    SIGNAL IMAGE AND VIDEO PROCESSING, 2024, 18 (SUPPL 1) : 31 - 41
  • [24] Multi-scale feature fusion network with local attention for lung segmentation
    Xie, Yinghua
    Zhou, Yuntong
    Wang, Chen
    Ma, Yanshan
    Yang, Ming
    SIGNAL PROCESSING-IMAGE COMMUNICATION, 2023, 119
  • [25] Text Detection Algorithm Based on Multi-Scale Attention Feature Fusion
    She, Xiangyang
    Liu, Zhe
    Dong, Lihong
    Computer Engineering and Applications, 2024, 60 (01) : 198 - 206
  • [26] Multi-Scale Feature Attention Fusion for Image Splicing Forgery Detection
    Liang, Enji
    Zhang, Kuiyuan
    Hua, Zhongyun
    Jia, Xiaohua
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2025, 21 (01)
  • [27] Small Object Detection using Multi-scale Feature Fusion and Attention
    Liu, Baokai
    Du, Shiqiang
    Li, Jiacheng
    Wang, Jianhua
    Liu, Wenjie
    2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 7246 - 7251
  • [28] Pyramid attention object detection network with multi-scale feature fusion
    Chen, Xiu
    Li, Yujie
    Nakatoh, Yoshihisa
    COMPUTERS & ELECTRICAL ENGINEERING, 2022, 104
  • [29] Adaptive feature fusion with attention mechanism for multi-scale target detection
    Moran Ju
    Jiangning Luo
    Zhongbo Wang
    Haibo Luo
    Neural Computing and Applications, 2021, 33 : 2769 - 2781
  • [30] Multi-Scale Feature Fusion Network with Attention for Single Image Dehazing
    Hu, Bin
    PATTERN RECOGNITION AND IMAGE ANALYSIS, 2021, 31 (04) : 608 - 615