Improved Multi-modal Image Fusion with Attention and Dense Networks: Visual and Quantitative Evaluation

被引:0
|
作者
Banerjee, Ankan [1 ]
Patra, Dipti [1 ]
Roy, Pradipta [2 ]
机构
[1] Natl Inst Technol, Rourkela, India
[2] DRDO, Integrated Test Range, Candipur, India
关键词
image fusion; attention; human perception; Convolutional Block Attention Module;
D O I
10.1007/978-3-031-58535-7_20
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This article introduces a novel multi-modal image fusion approach based on Convolutional Block Attention Module and dense networks to enhance human perceptual quality and information content in the fused images. The proposed model preserves the edges of the infrared images and enhances the contrast of the visible image as a pre-processing part. Consequently, the use of Convolutional Block Attention Module has resulted in the extraction of more refined features from the source images. The visual results demonstrate that the fused images produced by the proposed method are visually superior to those generated by most standard fusion techniques. To substantiate the findings, quantitative analysis is conducted using various metrics. The proposed method exhibits the best Naturalness Image Quality Evaluator and Chen-Varshney metric values, which are human perception-based parameters. Moreover, the fused images exhibit the highest Standard Deviation value, signifying enhanced contrast. These results justify the proposed multi-modal image fusion technique outperforms standard methods both qualitatively and quantitatively, resulting in superior fused images with improved human perception quality.
引用
收藏
页码:237 / 248
页数:12
相关论文
共 50 条
  • [21] Colour image cross-modal retrieval method based on multi-modal visual data fusion
    Liu, Xiangyuan
    International Journal of Computational Intelligence Studies, 2023, 12 (1-2) : 118 - 129
  • [22] Fusion of Multi-Modal Features to Enhance Dense Video Caption
    Huang, Xuefei
    Chan, Ka-Hou
    Wu, Weifan
    Sheng, Hao
    Ke, Wei
    SENSORS, 2023, 23 (12)
  • [23] Guided Image Deblurring by Deep Multi-Modal Image Fusion
    Liu, Yuqi
    Sheng, Zehua
    Shen, Hui-Liang
    IEEE ACCESS, 2022, 10 : 130708 - 130718
  • [24] Image Visual Attention Mechanism-based Global and Local Semantic Information Fusion for Multi-modal English Machine Translation
    Zhengzhou Railway Vocational and Technical College, Zhengzhou
    450000, China
    J. Comput., 2 (37-50): : 37 - 50
  • [25] Multi-modal Fusion
    Liu, Huaping
    Hussain, Amir
    Wang, Shuliang
    INFORMATION SCIENCES, 2018, 432 : 462 - 462
  • [26] Multi-modal medical image fusion using improved dual-channel PCNN
    Sinha, Adarsh
    Agarwal, Rahul
    Kumar, Vinay
    Garg, Nitin
    Pundir, Dhruv Singh
    Singh, Harsimran
    Rani, Ritu
    Panigrahy, Chinmaya
    MEDICAL & BIOLOGICAL ENGINEERING & COMPUTING, 2024, 62 (09) : 2629 - 2651
  • [27] Multi-modal feature fusion for geographic image annotation
    Li, Ke
    Zou, Changqing
    Bu, Shuhui
    Liang, Yun
    Zhang, Jian
    Gong, Minglun
    PATTERN RECOGNITION, 2018, 73 : 1 - 14
  • [28] A novel multi-modal medical image fusion algorithm
    Xinhua Li
    Jing Zhao
    Journal of Ambient Intelligence and Humanized Computing, 2021, 12 : 1995 - 2002
  • [29] Image and Encoded Text Fusion for Multi-Modal Classification
    Gallo, I.
    Calefati, A.
    Nawaz, S.
    Janjua, M. K.
    2018 INTERNATIONAL CONFERENCE ON DIGITAL IMAGE COMPUTING: TECHNIQUES AND APPLICATIONS (DICTA), 2018, : 203 - 209
  • [30] A novel multi-modal medical image fusion algorithm
    Li, Xinhua
    Zhao, Jing
    JOURNAL OF AMBIENT INTELLIGENCE AND HUMANIZED COMPUTING, 2021, 12 (02) : 1995 - 2002