Two-stage Unidirectional Fusion Network for RGBT tracking

被引:0
|
作者
Liu, Yisong [1 ]
Gao, Zhao [1 ]
Cao, Yang [2 ]
Zhou, Dongming [1 ,3 ]
机构
[1] Yunnan Univ, Sch Informat Sci & Engn, Kunming 650504, Yunnan, Peoples R China
[2] Southeast Univ, Sch Cyber Sci & Engn, Nanjing 210096, Peoples R China
[3] Hunan Univ Informat Technol, Sch Elect Sci & Engn, Changsha 410100, Peoples R China
基金
中国国家自然科学基金;
关键词
RGBT object tracking; Prompt learning; Multi-modal fusion; Causal decoder;
D O I
10.1016/j.knosys.2025.112983
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
RGB and Thermal (RGBT) tracking has recently attracted significant attention for its ability to accurately localize targets in complex scenarios. However, the creation of large-scale RGBT tracking datasets is both resource-intensive and laborious, motivating researchers to develop prompt tuning methods to adapt upstream RGB trackers to multimodal data with minimal additional parameters. Nevertheless, these methods do not fully exploit the supplementary modality information and tend to overlook the dynamic advantages between the two modalities in challenging scenarios. To address these issues, we propose a Two-stage Unidirectional Fusion (TUF) algorithm for RGBT tracking. This approach maximizes knowledge retention from upstream models while effectively leveraging the complementarity between the two modalities. It allows the powerful RGB feature extraction backbone from the upstream model to guide TIR image feature extraction through a two-stage unidirectional fusion strategy. Additionally, we have introduced an autoregressive decoder into RGBT tracking as a replacement for traditional bounding box prediction heads. This streamlines the framework of our RGBT tracker and improves tracking accuracy. Extensive experiments conducted on four widely used RGBT tracking benchmarks validate that our method surpasses existing state-of-the-art prompt tuning approaches, achieving a superior balance between performance and efficiency.
引用
收藏
页数:13
相关论文
共 50 条
  • [41] A two-stage algorithm for network reconstruction
    Wu, Jianshe
    Yang, Hedi
    Ren, Yanhua
    Li, X. Rong
    APPLIED SOFT COMPUTING, 2018, 70 : 751 - 763
  • [42] Two-stage wireless network emulation
    Pérennou, T
    Conchon, E
    Dairaine, L
    Diaz, M
    BROADBAND SATELLITE COMMUNICATION SYSTEMS AND THE CHALLENGES OF MOBILITY, 2005, 169 : 181 - 190
  • [43] Differential Reinforcement and Global Collaboration Network for RGBT Tracking
    Mei, Jiatian
    Zhou, Dongming
    Cao, Jinde
    Nie, Rencan
    He, Kangjian
    IEEE SENSORS JOURNAL, 2023, 23 (07) : 7301 - 7311
  • [44] TSFNet: Two-Stage Fusion Network for RGB-T Salient Object Detection
    Guo, Qinling
    Zhou, Wujie
    Lei, Jingsheng
    Yu, Lu
    IEEE SIGNAL PROCESSING LETTERS, 2021, 28 : 1655 - 1659
  • [45] A Two-Stage Special Feature Deep Fusion Network with Spatial Attention for Hippocampus Segmentation
    Cai, Zhengwei
    Wang, Shaoyu
    Chen, Qiang
    Lin, Runlong
    Hu, Yun
    Zhu, Yian
    2021 IEEE INTERNATIONAL CONFERENCE ON INFORMATION COMMUNICATION AND SOFTWARE ENGINEERING (ICICSE 2021), 2021, : 103 - 106
  • [46] Two-stage progressive residual learning network for multi-focus image fusion
    Wang, Haoran
    Hua, Zhen
    Li, Jinjiang
    IET IMAGE PROCESSING, 2022, 16 (03) : 772 - 786
  • [47] A pan-sharpening network using multiresolution transformer and two-stage feature fusion
    Fan W.
    Liu F.
    Li J.
    PeerJ Computer Science, 2023, 9
  • [48] Two-Stage Focus Measurement Network with Joint Boundary Refinement for Multifocus Image Fusion
    Zhai, Hao
    Pan, Xin
    Yang, You
    Jiang, Jinyuan
    Li, Qing
    INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2023, 2023
  • [49] An Efficient Two-Stage Fusion Network for Computer-Aided Diagnosis of Diabetic Foot
    Song, Anping
    Zhu, Hongtao
    Liu, Lifang
    Song, Ziheng
    Jin, Hongyu
    BIOINFORMATICS RESEARCH AND APPLICATIONS, ISBRA 2021, 2021, 13064 : 118 - 129
  • [50] Specific and Collaborative Representations Siamese Network for RGBT Tracking
    Liu, Yisong
    Zhou, Dongming
    Cao, Jinde
    Yan, Kaixiang
    Geng, Lizhi
    IEEE SENSORS JOURNAL, 2024, 24 (11) : 18520 - 18534