Two-stage Unidirectional Fusion Network for RGBT tracking

被引:0
|
作者
Liu, Yisong [1 ]
Gao, Zhao [1 ]
Cao, Yang [2 ]
Zhou, Dongming [1 ,3 ]
机构
[1] Yunnan Univ, Sch Informat Sci & Engn, Kunming 650504, Yunnan, Peoples R China
[2] Southeast Univ, Sch Cyber Sci & Engn, Nanjing 210096, Peoples R China
[3] Hunan Univ Informat Technol, Sch Elect Sci & Engn, Changsha 410100, Peoples R China
基金
中国国家自然科学基金;
关键词
RGBT object tracking; Prompt learning; Multi-modal fusion; Causal decoder;
D O I
10.1016/j.knosys.2025.112983
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
RGB and Thermal (RGBT) tracking has recently attracted significant attention for its ability to accurately localize targets in complex scenarios. However, the creation of large-scale RGBT tracking datasets is both resource-intensive and laborious, motivating researchers to develop prompt tuning methods to adapt upstream RGB trackers to multimodal data with minimal additional parameters. Nevertheless, these methods do not fully exploit the supplementary modality information and tend to overlook the dynamic advantages between the two modalities in challenging scenarios. To address these issues, we propose a Two-stage Unidirectional Fusion (TUF) algorithm for RGBT tracking. This approach maximizes knowledge retention from upstream models while effectively leveraging the complementarity between the two modalities. It allows the powerful RGB feature extraction backbone from the upstream model to guide TIR image feature extraction through a two-stage unidirectional fusion strategy. Additionally, we have introduced an autoregressive decoder into RGBT tracking as a replacement for traditional bounding box prediction heads. This streamlines the framework of our RGBT tracker and improves tracking accuracy. Extensive experiments conducted on four widely used RGBT tracking benchmarks validate that our method surpasses existing state-of-the-art prompt tuning approaches, achieving a superior balance between performance and efficiency.
引用
收藏
页数:13
相关论文
共 50 条
  • [31] Two-stage Tracking Algorithm for Passive Radar
    Malanowski, Mateusz
    Kulpa, Krzysztof
    Suchozebrski, Radoslaw
    FUSION: 2009 12TH INTERNATIONAL CONFERENCE ON INFORMATION FUSION, VOLS 1-4, 2009, : 1800 - 1806
  • [32] Two-Stage Saliency Fusion for Object Segmentation
    Sun, Guangling
    Ren, Jingru
    Liu, Zhi
    Shan, Wei
    IMAGE AND GRAPHICS (ICIG 2017), PT I, 2017, 10666 : 139 - 148
  • [33] MCSSAFNet: A multi-scale state-space attention fusion network for RGBT tracking
    Zhao, Chunbo
    Mo, Bo
    Li, Dawei
    Wang, Xinchun
    Zhao, Jie
    Xu, Junwei
    OPTICS COMMUNICATIONS, 2025, 577
  • [34] A Two-Stage Network for Image Deblurring
    Pan, Ze
    Lv, Qunbo
    Tan, Zheng
    IEEE ACCESS, 2021, 9 : 76707 - 76715
  • [35] Robust Two-Stage Network Problems
    Kasperski, Adam
    Zielinski, Pawel
    OPERATIONS RESEARCH PROCEEDINGS 2015, 2017, : 35 - 40
  • [36] A Two-Stage Approach for Network Monitoring
    Bai, Linda
    Roy, Sumit
    JOURNAL OF NETWORK AND SYSTEMS MANAGEMENT, 2013, 21 (02) : 238 - 263
  • [37] A Two-Stage Approach for Network Monitoring
    Linda Bai
    Sumit Roy
    Journal of Network and Systems Management, 2013, 21 : 238 - 263
  • [38] Exploring fusion strategies for accurate RGBT visual object tracking
    Tang, Zhangyong
    Xu, Tianyang
    Li, Hui
    Wu, Xiao-Jun
    Zhu, XueFeng
    Kittler, Josef
    INFORMATION FUSION, 2023, 99
  • [39] Two-stage multi-radar data fusion for re-entry target tracking
    Jung B.
    Lee C.-S.
    Suh U.-S.
    Ra W.-S.
    Transactions of the Korean Institute of Electrical Engineers, 2021, 70 (09): : 1372 - 1380
  • [40] Analysis of a two-stage network server
    Salah, K.
    APPLIED MATHEMATICS AND COMPUTATION, 2011, 217 (23) : 9635 - 9645