DACFusion: Dual Asymmetric Cross-Attention guided feature fusion for multispectral object detection

被引:0
|
作者
Qian, Jingchen [1 ]
Qiao, Baiyou [1 ,2 ]
Zhang, Yuekai [1 ]
Liu, Tongyan [1 ]
Wang, Shuo [1 ]
Wu, Gang [1 ,2 ]
Han, Donghong [1 ,2 ]
机构
[1] Northeastern Univ, Sch Comp Sci & Engn, Shenyang 110819, Peoples R China
[2] Northeastern Univ, Key Lab Intelligent Comp Med Image, Minist Educ, Shenyang 110819, Peoples R China
关键词
Multispectral object detection; Cross-attention; Feature fusion; SCALING-UP; NETWORK;
D O I
10.1016/j.neucom.2025.129913
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Effective fusion of unique features from different spectra plays a crucial role in multispectral object detection. Recent research has focused on transplanting advanced methods from other multimodal fusion fields to multispectral object detection tasks. These fusion methods focus on the fusion of features and ignore the spatial correspondence between multispectral images. This lack of correspondence in turn limits the full utilization of the complementarities between the different modalities, which affects the accuracy of object detection. To address this problem, we creatively propose a dual asymmetric cross-attention multispectral fusion (DACFusion) method, which is able to process features interactively based on the positional correspondence between two spectra, and then asymmetrically fuses the multispectral data according to the characteristics of each spectrum to take advantage of their complementary strengths. Meanwhile, we introduce a large selective kernel network to expand the receptive field for object detection, which further improves the detection accuracy. Experimental results on the VEDAI and LLVIP datasets validate the significant performance advantages of our proposed method and show its applicability to a variety of practical application scenarios. Code will be available at https://github.com/wood-fish/DACFusion.
引用
收藏
页数:13
相关论文
共 50 条
  • [1] ICAFusion: Iterative cross-attention guided feature fusion for multispectral object detection
    Shen, Jifeng
    Chen, Yifei
    Liu, Yue
    Zuo, Xin
    Fan, Heng
    Yang, Wankou
    PATTERN RECOGNITION, 2024, 145
  • [2] Background-Aware Cross-Attention Multiscale Fusion for Multispectral Object Detection
    Guo, Runze
    Guo, Xiaojun
    Sun, Xiaoyong
    Zhou, Peida
    Sun, Bei
    Su, Shaojing
    REMOTE SENSING, 2024, 16 (21)
  • [3] Object Detection in Multispectral Remote Sensing Images Based on Cross-Modal Cross-Attention
    Zhao, Pujie
    Ye, Xia
    Du, Ziang
    SENSORS, 2024, 24 (13)
  • [4] Multispectral Object Detection Based on Multilevel Feature Fusion and Dual Feature Modulation
    Sun, Jin
    Yin, Mingfeng
    Wang, Zhiwei
    Xie, Tao
    Bei, Shaoyi
    ELECTRONICS, 2024, 13 (02)
  • [5] Object Detection by Attention-Guided Feature Fusion Network
    Shi, Yuxuan
    Fan, Yue
    Xu, Siqi
    Gao, Yue
    Gao, Ran
    SYMMETRY-BASEL, 2022, 14 (05):
  • [6] Attention-guided Feature Fusion for Small Object Detection
    Yang, Jiaxiong
    Liu, Xianhui
    Liu, Zhuang
    IST 2023 - IEEE International Conference on Imaging Systems and Techniques, Proceedings, 2023,
  • [7] YOLO-MS: Multispectral Object Detection via Feature Interaction and Self-Attention Guided Fusion
    Xie, Yumin
    Zhang, Langwen
    Yu, Xiaoyuan
    Xie, Wei
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2023, 15 (04) : 2132 - 2143
  • [8] Spatial Cross-Attention RGB-D Fusion Module for Object Detection
    Gao, Shangyin
    Markhasin, Lev
    Wang, Bi
    IEEE MMSP 2021: 2021 IEEE 23RD INTERNATIONAL WORKSHOP ON MULTIMEDIA SIGNAL PROCESSING (MMSP), 2021,
  • [9] DyFusion: Cross-Attention 3D Object Detection with Dynamic Fusion
    Bi, Jiangfeng
    Wei, Haiyue
    Zhang, Guoxin
    Yang, Kuihe
    Song, Ziying
    IEEE LATIN AMERICA TRANSACTIONS, 2024, 22 (02) : 106 - 112
  • [10] Learnable Cross-Scale Sparse Attention Guided Feature Fusion for UAV Object Detection
    Zuo, Xin
    Qi, Chenhui
    Chen, Yifei
    Shen, Jifeng
    Fan, Heng
    Yang, Wankou
    IEEE ACCESS, 2024, 12 : 114212 - 114226