AFSFusion: An Adjacent Feature Shuffle Combination Network for Infrared and Visible Image Fusion

被引:1
|
作者
Hu, Yufeng [1 ]
Xu, Shaoping [2 ]
Cheng, Xiaohui [2 ]
Zhou, Changfei [2 ]
Xiong, Minghai [2 ]
机构
[1] Nanchang Univ, Sch Qianhu, Nanchang 330031, Peoples R China
[2] Nanchang Univ, Sch Math & Comp Sci, Nanchang 330031, Peoples R China
来源
APPLIED SCIENCES-BASEL | 2023年 / 13卷 / 09期
关键词
infrared and visible image fusion; adjacent feature shuffle fusion; adaptive weight adjustment strategy; subjective and objective evaluation; INFORMATION;
D O I
10.3390/app13095640
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
To obtain fused images with excellent contrast, distinct target edges, and well-preserved details, we propose an adaptive image fusion network called the adjacent feature shuffle-fusion network (AFSFusion). The proposed network adopts a UNet-like architecture and incorporates key refinements to enhance network architecture and loss functions. Regarding the network architecture, the proposed two-branch adjacent feature fusion module, called AFSF, expands the number of channels to fuse the feature channels of several adjacent convolutional layers in the first half of the AFSFusion, enhancing its ability to extract, transmit, and modulate feature information. We replace the original rectified linear unit (ReLU) with leaky ReLU to alleviate the problem of gradient disappearance and add a channel shuffling operation at the end of AFSF to facilitate information interaction capability between features. Concerning loss functions, we propose an adaptive weight adjustment (AWA) strategy to assign weight values to the corresponding pixels of the infrared (IR) and visible images in the fused images, according to the VGG16 gradient feature response of the IR and visible images. This strategy efficiently handles different scene contents. After normalization, the weight values are used as weighting coefficients for the two sets of images. The weighting coefficients are applied to three loss items simultaneously: mean square error (MSE), structural similarity (SSIM), and total variation (TV), resulting in clearer objects and richer texture detail in the fused images. We conducted a series of experiments on several benchmark databases, and the results demonstrate the effectiveness of the proposed network architecture and the superiority of the proposed network compared to other state-of-the-art fusion methods. It ranks first in several objective metrics, showing the best performance and exhibiting sharper and richer edges of specific targets, which is more in line with human visual perception. The remarkable enhancement in performance is ascribed to the proposed AFSF module and AWA strategy, enabling balanced feature extraction, fusion, and modulation of image features throughout the process.
引用
收藏
页数:20
相关论文
共 50 条
  • [1] Adjacent Feature Combination Based Adaptive Fusion Network for Infrared and Visible Images
    Xu S.
    Chen X.
    Luo J.
    Cheng X.
    Xiao N.
    Moshi Shibie yu Rengong Zhineng/Pattern Recognition and Artificial Intelligence, 2022, 35 (12): : 1089 - 1100
  • [2] An Information Retention and Feature Transmission Network for Infrared and Visible Image Fusion
    Liu, Chang
    Yang, Bin
    Li, Yuehua
    Zhang, Xiaozhi
    Pang, Lihui
    IEEE SENSORS JOURNAL, 2021, 21 (13) : 14950 - 14959
  • [3] FPNFuse: A lightweight feature pyramid network for infrared and visible image fusion
    Zhang, Zi-Han
    Wu, Xiao-Jun
    Xu, Tianyang
    IET IMAGE PROCESSING, 2022, 16 (09) : 2308 - 2320
  • [4] BDPartNet: Feature Decoupling and Reconstruction Fusion Network for Infrared and Visible Image
    Wang, Xuejie
    Zhang, Jianxun
    Tao, Ye
    Yuan, Xiaoli
    Guo, Yifan
    CMC-COMPUTERS MATERIALS & CONTINUA, 2024, 79 (03): : 4621 - 4639
  • [5] Interactive Feature Embedding for Infrared and Visible Image Fusion
    Zhao, Fan
    Zhao, Wenda
    Lu, Huchuan
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (09) : 12810 - 12822
  • [6] Infrared and Visible Image Fusion Based on Sparse Feature
    Ding Wen-shan
    Bi Du-yan
    He Lin-yuan
    Fan Zun-lin
    Wu Dong-peng
    ACTA PHOTONICA SINICA, 2018, 47 (09)
  • [7] Dual-Attention-Based Feature Aggregation Network for Infrared and Visible Image Fusion
    Tang, Zhimin
    Xiao, Guobao
    Guo, Junwen
    Wang, Shiping
    Ma, Jiayi
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2023, 72
  • [8] DBIF: Dual-Branch Feature Extraction Network for Infrared and Visible Image Fusion
    Zhang, Haozhe
    Cui, Rongpu
    Zheng, Zhuohang
    Gao, Shaobing
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2024, PT VIII, 2025, 15038 : 309 - 323
  • [9] THFuse: An infrared and visible image fusion network using transformer and hybrid feature extractor
    Chen, Jun
    Ding, Jianfeng
    Yu, Yang
    Gong, Wenping
    NEUROCOMPUTING, 2023, 527 : 71 - 82
  • [10] FTSFN: A Two-Stage Feature Transfer and Supplement Fusion Network for Infrared and Visible Image Fusion
    Huang, Shuying
    Kong, Xiangkai
    Yang, Yong
    Wan, Weiguo
    Song, Zixiang
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2025, 74