Semantic attention-based heterogeneous feature aggregation network for image fusion

被引:1
|
作者
Ruan, Zhiqiang [1 ]
Wan, Jie [1 ,2 ]
Xiao, Guobao [3 ]
Tang, Zhimin [2 ]
Ma, Jiayi [4 ]
机构
[1] Minjiang Univ, Coll Comp & Data Sci, Fuzhou 350108, Peoples R China
[2] Fuzhou Univ, Coll Comp & Data Sci, Fuzhou 350108, Peoples R China
[3] Tongji Univ, Sch Elect & Informat Engn, Shanghai 200092, Peoples R China
[4] Wuhan Univ, Elect Informat Sch, Wuhan 430072, Peoples R China
关键词
Image fusion; High-level vision tasks; Attention mechanism; Semantic prior;
D O I
10.1016/j.patcog.2024.110728
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Infrared and visible image fusion aims to generate a comprehensive image that retains both salient targets of the infrared image and texture details of the visible image. However, existing methods overlook the differences in attention to semantic information among different fused images. To address this issue, we propose a semantic attention-based heterogeneous feature aggregation network for image fusion. The key component of our network is the semantic attention-based fusion module, which leverages the weights derived from semantic feature maps to dynamically adjust the significance of various semantic objects within the fusion feature maps. By using semantic weights as guidance, our fusion process concentrates on regions with crucial semantics, resulting in a more focused fusion that preserves rich semantic information. Moreover, we propose an innovative component called the attentive dense block. This block effectively filters out irrelevant features during extraction, accentuates essential features to their maximum potential, and enhances the visual quality of the fused images. Importantly, our network demonstrates strong generalization capabilities. Extensive experiments validate the superiority of our proposed network over current state-of-the-art techniques in terms of both visual appeal and semantics-driven evaluation.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Bidirectional Feature Aggregation Network for Stereo Image Quality Assessment Considering Parallax Attention-Based Binocular Fusion
    Chang, Yongli
    Li, Sumei
    Liu, Anqi
    Zhang, Wenlin
    Jin, Jie
    Xiang, Wei
    IEEE TRANSACTIONS ON BROADCASTING, 2024, 70 (01) : 278 - 289
  • [2] Attention-based dual context aggregation for image semantic segmentation
    Zhao, Dexin
    Qi, Zhiyang
    Yang, Ruixue
    Wang, Zhaohui
    MULTIMEDIA TOOLS AND APPLICATIONS, 2021, 80 (18) : 28201 - 28216
  • [3] Attention-based dual context aggregation for image semantic segmentation
    Dexin Zhao
    Zhiyang Qi
    Ruixue Yang
    Zhaohui Wang
    Multimedia Tools and Applications, 2021, 80 : 28201 - 28216
  • [4] Semantic Correlation Attention-Based Multiorder Multiscale Feature Fusion Network for Human Motion Prediction
    Li, Qin
    Wang, Yong
    Lv, Fanbing
    IEEE TRANSACTIONS ON CYBERNETICS, 2024, 54 (02) : 825 - 838
  • [5] Attention-based Fusion Network for Image Forgery Localization
    Gong, Wenhui
    Chen, Yan
    Alam, Mohammad S.
    Sang, Jun
    PATTERN RECOGNITION AND PREDICTION XXXV, 2024, 13040
  • [6] Attention-based acoustic feature fusion network for depression detection
    Xu, Xiao
    Wang, Yang
    Wei, Xinru
    Wang, Fei
    Zhang, Xizhe
    NEUROCOMPUTING, 2024, 601
  • [7] Infrared and Visible Image Fusion via Attention-Based Adaptive Feature Fusion
    Wang, Lei
    Hu, Ziming
    Kong, Quan
    Qi, Qian
    Liao, Qing
    ENTROPY, 2023, 25 (03)
  • [8] Dual-Attention-Based Feature Aggregation Network for Infrared and Visible Image Fusion
    Tang, Zhimin
    Xiao, Guobao
    Guo, Junwen
    Wang, Shiping
    Ma, Jiayi
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2023, 72
  • [9] Attention-based fusion network for RGB-D semantic segmentation
    Zhong, Li
    Guo, Chi
    Zhan, Jiao
    Deng, JingYi
    NEUROCOMPUTING, 2024, 608
  • [10] Meta network attention-based feature matching for heterogeneous defect prediction
    Nevendra, Meetesh
    Singh, Pradeep
    AUTOMATED SOFTWARE ENGINEERING, 2025, 32 (01)