GLMAFuse: A Dual-Stream Infrared and Visible Image Fusion Framework Integrating Local and Global Features with Multi-Scale Attention

被引:0
|
作者
Li, Fu [1 ,2 ,3 ]
Gu, Yanghai [4 ]
Zhao, Ming [1 ]
Chen, Deji [1 ,3 ]
Wang, Quan [1 ]
机构
[1] Wuxi Univ, Sch Internet Things Engn, Wuxi 214105, Peoples R China
[2] Wuxi Univ, Jiangsu Engn Res Ctr Hyperconvergence Applicat & S, Wuxi 214105, Peoples R China
[3] Tongji Univ, Minist Educ, Key Lab Embedded Syst & Serv Comp, Shanghai 201804, Peoples R China
[4] Nanjing Univ Informat Sci & Technol, Sch Comp Sci & Technol, Nanjing 210044, Peoples R China
来源
ELECTRONICS | 2024年 / 13卷 / 24期
关键词
image fusion; global and local feature; multi-scale; dual-stream; attention mechanism; NETWORK; NEST;
D O I
10.3390/electronics13245002
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Integrating infrared and visible-light images facilitates a more comprehensive understanding of scenes by amalgamating dual-sensor data derived from identical environments. Traditional CNN-based fusion techniques are predominantly confined to local feature emphasis due to their inherently limited receptive fields. Conversely, Transformer-based models tend to prioritize global information, which can lead to a deficiency in feature diversity and detail retention. Furthermore, methods reliant on single-scale feature extraction are inadequate for capturing extensive scene information. To address these limitations, this study presents GLMAFuse, an innovative dual-stream encoder-decoder network, which utilizes a multi-scale attention mechanism to harmoniously integrate global and local features. This framework is designed to maximize the extraction of multi-scale features from source images while effectively synthesizing local and global information across all layers. We introduce the global-aware and local embedding (GALE) module to adeptly capture and merge global structural attributes and localized details from infrared and visible imagery via a parallel dual-branch architecture. Additionally, the multi-scale attention fusion (MSAF) module is engineered to optimize attention weights at the channel level, facilitating an enhanced synergy between high-frequency edge details and global backgrounds. This promotes effective interaction and fusion of dual-modal features. Extensive evaluations using standard datasets demonstrate that GLMAFuse surpasses the existing leading methods in both qualitative and quantitative assessments, highlighting its superior capability in infrared and visible image fusion. On the TNO and MSRS datasets, our method achieves outstanding performance across multiple metrics, including EN (7.15, 6.75), SD (46.72, 47.55), SF (12.79, 12.56), MI (2.21, 3.22), SCD (1.75, 1.80), VIF (0.79, 1.08), Qbaf (0.58, 0.71), and SSIM (0.99, 1.00). These results underscore its exceptional proficiency in infrared and visible image fusion.
引用
收藏
页数:30
相关论文
共 50 条
  • [1] Integrating Parallel Attention Mechanisms and Multi-Scale Features for Infrared and Visible Image Fusion
    Xu, Qian
    Zheng, Yuan
    IEEE ACCESS, 2024, 12 : 8359 - 8372
  • [2] Prompt learning and multi-scale attention for infrared and visible image fusion
    Li, Yanan
    Ji, Qingtao
    Jiao, Shaokang
    INFRARED PHYSICS & TECHNOLOGY, 2025, 145
  • [3] Multi-scale infrared and visible image fusion framework based on dual partial differential equations
    Guo, Chentong
    Liu, Chenhua
    Deng, Lei
    Chen, Zhixiang
    Dong, Mingli
    Zhu, Lianqing
    Chen, Hanrui
    Lu, Xitian
    INFRARED PHYSICS & TECHNOLOGY, 2023, 135
  • [4] A multi-scale information integration framework for infrared and visible image fusion
    Yang, Guang
    Li, Jie
    Lei, Hanxiao
    Gao, Xinbo
    NEUROCOMPUTING, 2024, 600
  • [5] Dual-Stream Fusion and Multi-scale Analysis: Introducing the Synergistic Dual-Stream Network (SDS-Net) for Image Manipulation Segmentation
    Song, Huaqing
    Lin, Baichuan
    Xie, Lanchi
    ADVANCED INTELLIGENT SYSTEMS, 2024, 6 (04)
  • [6] An infrared and visible image fusion network based on multi-scale feature cascades and non-local attention
    Xu, Jing
    Liu, Zhenjin
    Fang, Ming
    IET IMAGE PROCESSING, 2024, 18 (08) : 2114 - 2125
  • [7] Infrared and visible image fusion based on multi-scale dense attention connection network
    Chen Y.
    Zhang J.
    Wang Z.
    Guangxue Jingmi Gongcheng/Optics and Precision Engineering, 2022, 30 (18): : 2253 - 2266
  • [8] An Infrared and Visible Image Fusion Framework based on Dual Scale Decomposition and Learnable Attention Fusion Strategy
    Cheng, Guanzheng
    Jin, Lizuo
    Chai, Lin
    2023 35TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2023, : 4087 - 4092
  • [9] Multi-scale single image dehazing based on the fusion of global and local features
    Chen, Ziyu
    Zhuang, Huaiyu
    Han, Jia
    Cui, Yani
    Deng, Jiaxian
    IET IMAGE PROCESSING, 2022, 16 (08) : 2049 - 2062
  • [10] MFT: Multi-scale Fusion Transformer for Infrared and Visible Image Fusion
    Zhang, Chen-Ming
    Yuan, Chengbo
    Luo, Yong
    Zhou, Xin
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT VI, 2023, 14259 : 485 - 496