GLMAFuse: A Dual-Stream Infrared and Visible Image Fusion Framework Integrating Local and Global Features with Multi-Scale Attention

被引:0
|
作者
Li, Fu [1 ,2 ,3 ]
Gu, Yanghai [4 ]
Zhao, Ming [1 ]
Chen, Deji [1 ,3 ]
Wang, Quan [1 ]
机构
[1] Wuxi Univ, Sch Internet Things Engn, Wuxi 214105, Peoples R China
[2] Wuxi Univ, Jiangsu Engn Res Ctr Hyperconvergence Applicat & S, Wuxi 214105, Peoples R China
[3] Tongji Univ, Minist Educ, Key Lab Embedded Syst & Serv Comp, Shanghai 201804, Peoples R China
[4] Nanjing Univ Informat Sci & Technol, Sch Comp Sci & Technol, Nanjing 210044, Peoples R China
来源
ELECTRONICS | 2024年 / 13卷 / 24期
关键词
image fusion; global and local feature; multi-scale; dual-stream; attention mechanism; NETWORK; NEST;
D O I
10.3390/electronics13245002
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Integrating infrared and visible-light images facilitates a more comprehensive understanding of scenes by amalgamating dual-sensor data derived from identical environments. Traditional CNN-based fusion techniques are predominantly confined to local feature emphasis due to their inherently limited receptive fields. Conversely, Transformer-based models tend to prioritize global information, which can lead to a deficiency in feature diversity and detail retention. Furthermore, methods reliant on single-scale feature extraction are inadequate for capturing extensive scene information. To address these limitations, this study presents GLMAFuse, an innovative dual-stream encoder-decoder network, which utilizes a multi-scale attention mechanism to harmoniously integrate global and local features. This framework is designed to maximize the extraction of multi-scale features from source images while effectively synthesizing local and global information across all layers. We introduce the global-aware and local embedding (GALE) module to adeptly capture and merge global structural attributes and localized details from infrared and visible imagery via a parallel dual-branch architecture. Additionally, the multi-scale attention fusion (MSAF) module is engineered to optimize attention weights at the channel level, facilitating an enhanced synergy between high-frequency edge details and global backgrounds. This promotes effective interaction and fusion of dual-modal features. Extensive evaluations using standard datasets demonstrate that GLMAFuse surpasses the existing leading methods in both qualitative and quantitative assessments, highlighting its superior capability in infrared and visible image fusion. On the TNO and MSRS datasets, our method achieves outstanding performance across multiple metrics, including EN (7.15, 6.75), SD (46.72, 47.55), SF (12.79, 12.56), MI (2.21, 3.22), SCD (1.75, 1.80), VIF (0.79, 1.08), Qbaf (0.58, 0.71), and SSIM (0.99, 1.00). These results underscore its exceptional proficiency in infrared and visible image fusion.
引用
收藏
页数:30
相关论文
共 50 条
  • [31] MMF: A Multi-scale MobileNet based fusion method for infrared and visible image
    Liu, Yi
    Miao, Changyun
    Ji, Jianhua
    Li, Xianguo
    INFRARED PHYSICS & TECHNOLOGY, 2021, 119
  • [32] Multi-scale saliency measure and orthogonal space for visible and infrared image fusion
    Liu, Yaochen
    Dong, Lili
    Ren, Wei
    Xu, Wenhai
    INFRARED PHYSICS & TECHNOLOGY, 2021, 118
  • [33] Underwater Image Enhancement Based on Global and Local Equalization of Histogram and Dual-Image Multi-Scale Fusion
    Bai, Linfeng
    Zhang, Weidong
    Pan, Xipeng
    Zhao, Chenping
    IEEE ACCESS, 2020, 8 : 128973 - 128990
  • [34] Dual-stream network with complementary fusion and hierarchical attention for image tampering localization
    Mao, Zhanpeng
    Lu, Tongwei
    SIGNAL IMAGE AND VIDEO PROCESSING, 2025, 19 (03)
  • [35] Infrared and visible image fusion using multi-scale edge-preserving decomposition and multiple saliency features
    Duan, Chaowei
    Wang, Zhisheng
    Xing, Changda
    Lu, Shanshan
    OPTIK, 2021, 228
  • [36] Lightweight Image Super-Resolution Based on Local Interaction of Multi-Scale Features and Global Fusion
    Meng, Zhiqing
    Zhang, Jing
    Li, Xiangjun
    Zhang, Lingyin
    MATHEMATICS, 2022, 10 (07)
  • [37] Infrared and visible image fusion via saliency analysis and local edge-preserving multi-scale decomposition
    Zhang, Xiaoye
    Ma, Yong
    Fan, Fan
    Zhang, Ying
    Huang, Jun
    JOURNAL OF THE OPTICAL SOCIETY OF AMERICA A-OPTICS IMAGE SCIENCE AND VISION, 2017, 34 (08) : 1400 - 1410
  • [38] Multi-scale Fusion of Stretched Infrared and Visible Images
    Jia, Weibin
    Song, Zhihuan
    Li, Zhengguo
    SENSORS, 2022, 22 (17)
  • [39] MDSCNN: Remote Sensing Image Spatial-Spectral Fusion Method via Multi-Scale Dual-Stream Convolutional Neural Network
    Wang, Wenqing
    Jia, Fei
    Yang, Yifei
    Mu, Kunpeng
    Liu, Han
    REMOTE SENSING, 2024, 16 (19)
  • [40] Deep Neural Network for Infrared and Visible Image Fusion Based on Multi-scale Decomposition and Interactive Residual Coordinate Attention
    Zong, Sha
    Xie, Zhihua
    Li, Qiang
    Liu, Guodong
    ADVANCES IN NATURAL COMPUTATION, FUZZY SYSTEMS AND KNOWLEDGE DISCOVERY, ICNC-FSKD 2022, 2023, 153 : 254 - 262