GLMAFuse: A Dual-Stream Infrared and Visible Image Fusion Framework Integrating Local and Global Features with Multi-Scale Attention

被引:0
|
作者
Li, Fu [1 ,2 ,3 ]
Gu, Yanghai [4 ]
Zhao, Ming [1 ]
Chen, Deji [1 ,3 ]
Wang, Quan [1 ]
机构
[1] Wuxi Univ, Sch Internet Things Engn, Wuxi 214105, Peoples R China
[2] Wuxi Univ, Jiangsu Engn Res Ctr Hyperconvergence Applicat & S, Wuxi 214105, Peoples R China
[3] Tongji Univ, Minist Educ, Key Lab Embedded Syst & Serv Comp, Shanghai 201804, Peoples R China
[4] Nanjing Univ Informat Sci & Technol, Sch Comp Sci & Technol, Nanjing 210044, Peoples R China
来源
ELECTRONICS | 2024年 / 13卷 / 24期
关键词
image fusion; global and local feature; multi-scale; dual-stream; attention mechanism; NETWORK; NEST;
D O I
10.3390/electronics13245002
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Integrating infrared and visible-light images facilitates a more comprehensive understanding of scenes by amalgamating dual-sensor data derived from identical environments. Traditional CNN-based fusion techniques are predominantly confined to local feature emphasis due to their inherently limited receptive fields. Conversely, Transformer-based models tend to prioritize global information, which can lead to a deficiency in feature diversity and detail retention. Furthermore, methods reliant on single-scale feature extraction are inadequate for capturing extensive scene information. To address these limitations, this study presents GLMAFuse, an innovative dual-stream encoder-decoder network, which utilizes a multi-scale attention mechanism to harmoniously integrate global and local features. This framework is designed to maximize the extraction of multi-scale features from source images while effectively synthesizing local and global information across all layers. We introduce the global-aware and local embedding (GALE) module to adeptly capture and merge global structural attributes and localized details from infrared and visible imagery via a parallel dual-branch architecture. Additionally, the multi-scale attention fusion (MSAF) module is engineered to optimize attention weights at the channel level, facilitating an enhanced synergy between high-frequency edge details and global backgrounds. This promotes effective interaction and fusion of dual-modal features. Extensive evaluations using standard datasets demonstrate that GLMAFuse surpasses the existing leading methods in both qualitative and quantitative assessments, highlighting its superior capability in infrared and visible image fusion. On the TNO and MSRS datasets, our method achieves outstanding performance across multiple metrics, including EN (7.15, 6.75), SD (46.72, 47.55), SF (12.79, 12.56), MI (2.21, 3.22), SCD (1.75, 1.80), VIF (0.79, 1.08), Qbaf (0.58, 0.71), and SSIM (0.99, 1.00). These results underscore its exceptional proficiency in infrared and visible image fusion.
引用
收藏
页数:30
相关论文
共 50 条
  • [21] Multi-scale attention-based lightweight network with dilated convolutions for infrared and visible image fusion
    Li, Fuquan
    Zhou, Yonghui
    Chen, YanLi
    Li, Jie
    Dong, ZhiCheng
    Tan, Mian
    COMPLEX & INTELLIGENT SYSTEMS, 2024, 10 (01) : 705 - 719
  • [22] Multi-scale attention-based lightweight network with dilated convolutions for infrared and visible image fusion
    Fuquan Li
    Yonghui Zhou
    YanLi Chen
    Jie Li
    ZhiCheng Dong
    Mian Tan
    Complex & Intelligent Systems, 2024, 10 : 705 - 719
  • [23] Infrared and visible image features enhancement and fusion using multi-scale top-hat decomposition
    Li, Yufeng
    Feng, Xiaoyun
    Xu, Mingwei
    Hongwai yu Jiguang Gongcheng/Infrared and Laser Engineering, 2012, 41 (10): : 2824 - 2832
  • [24] Infrared and Visible Image Fusion Based on Multi-scale Network with Dual-channel Information Cross Fusion Block
    Yang, Yong
    Kong, Xiangkai
    Huang, Shuying
    Wan, Weiguo
    Liu, Jiaxiang
    Zhang, Wang
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [25] A VISIBLE AND INFRARED IMAGE FUSION FRAMEWORK BASED ON DUAL-PATH ENCODER-DECODER AND MULTI-SCALE DISCRETE WAVELET TRANSFORM
    Liu, Renhe
    Wang, Han
    Du, Shan
    Liu, Yu
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 1995 - 1999
  • [26] MIAFusion: Infrared and Visible Image Fusion via Multi-scale Spatial and Channel-Aware Interaction Attention
    Lin, Teng
    Lu, Ming
    Jiang, Min
    Kong, Jun
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2024, PT VIII, 2025, 15038 : 238 - 251
  • [27] MRASFusion: A multi-scale residual attention infrared and visible image fusion network based on semantic segmentation guidance
    An, Rongsheng
    Liu, Gang
    Qian, Yao
    Xing, Mengliang
    Tang, Haojie
    INFRARED PHYSICS & TECHNOLOGY, 2024, 139
  • [28] MCADFusion: a novel multi-scale convolutional attention decomposition method for enhanced infrared and visible light image fusion
    Zhang, Wangwei
    Dai, Menghao
    Zhou, Bin
    Wang, Changhai
    ELECTRONIC RESEARCH ARCHIVE, 2024, 32 (08): : 5067 - 5089
  • [29] Infrared image fault diagnosis based on dual-stream attention convolution network
    Lu, Dong
    Yang, Jing
    Ming, Lyu
    Zhang, Jie
    ENGINEERING RESEARCH EXPRESS, 2024, 6 (02):
  • [30] A Multi-Scale Infrared and Visible Image Fusion Network Based on Context Perception
    Zhao, Huixuan
    Cheng, Jinyong
    Du, Rundong
    PROCEEDINGS OF THE 2024 27 TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, CSCWD 2024, 2024, : 395 - 400