CMFA_Net: A cross-modal feature aggregation network for infrared-visible image fusion

被引:15
|
作者
Ding, Zhaisheng [1 ]
Li, Haiyan [1 ]
Zhou, Dongming [1 ]
Li, Hongsong [1 ]
Liu, Yanyu [1 ]
Hou, Ruichao [2 ]
机构
[1] Yunnan Univ, Sch Informat, Kunming 650504, Yunnan, Peoples R China
[2] Nanjing Univ, State Key Lab Novel Software Technol, Nanjing 210023, Peoples R China
基金
中国国家自然科学基金;
关键词
Cross-modal; Attention mechanism; Image fusion; Unsupervised learning; End-to-end network; Infrared-visible images; PERFORMANCE; FRAMEWORK;
D O I
10.1016/j.infrared.2021.103905
中图分类号
TH7 [仪器、仪表];
学科分类号
0804 ; 080401 ; 081102 ;
摘要
Infrared and visible image fusion is a typical cross-modal information enhancement technology, which aims to fetch the complementary cues from different sensors to reconstruct an informative image or video. Many related works focus on designing the hand-crafted fusion rules and ignore the inner complementarity potentials of modalities, resulting in failing to mine the ability of the deep model thoroughly. In this work, an unsupervised cross-modal feature aggregation network (CMFA_Net) is developed, which explores the latent correlations between the internal characteristics effectively and processes these information to fuse a satisfactory image. Firstly, a densely integrated structure and an attention module are proposed to form a feature extractor. Subsequently, the l1 -norm and the attention mechanism are combined to fuse the affinity features of the cross-modal images. Finally, the fused image is reconstructed by the deconvolution block. To guarantee the clarity and rich information of the fused image, a specific loss function is put forward by utilizing the average pixel decision for structural similarity (SSIM-p) and content-gram variation (CGV) for training the model on the KAIST dataset. Extensive and solid experiments verify the effectiveness and robustness of the proposed model and demonstrate that the proposed method outperforms the state-of-the-arts and achieves advanced performance as well as less computational consumption both in subjective and objective evaluations.
引用
收藏
页数:13
相关论文
共 50 条
  • [1] Cross-Modal Transformers for Infrared and Visible Image Fusion
    Park, Seonghyun
    Vien, An Gia
    Lee, Chul
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (02) : 770 - 785
  • [2] BCMFIFuse: A Bilateral Cross-Modal Feature Interaction-Based Network for Infrared and Visible Image Fusion
    Gao, Xueyan
    Liu, Shiguang
    REMOTE SENSING, 2024, 16 (17)
  • [3] CCAFusion: Cross-Modal Coordinate Attention Network for Infrared and Visible Image Fusion
    Li, Xiaoling
    Li, Yanfeng
    Chen, Houjin
    Peng, Yahui
    Pan, Pan
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (02) : 866 - 881
  • [4] Infrared and visible image fusion based on cross-modal extraction strategy
    Liu, Xiaowen
    Li, Jing
    Yang, Xin
    Huo, Hongtao
    INFRARED PHYSICS & TECHNOLOGY, 2022, 124
  • [5] Efficient multi-level cross-modal fusion and detection network for infrared and visible image
    Gao, Hongwei
    Wang, Yutong
    Sun, Jian
    Jiang, Yueqiu
    Gai, Yonggang
    Yu, Jiahui
    ALEXANDRIA ENGINEERING JOURNAL, 2024, 108 : 306 - 318
  • [6] Infrared-Visible Cross-Modal Person Re-Identification with an X Modality
    Li, Diangang
    Wei, Xing
    Hong, Xiaopeng
    Gong, Yihong
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 4610 - 4617
  • [7] PFCFuse: A Poolformer and CNN Fusion Network for Infrared-Visible Image Fusion
    Hu, Xinyu
    Liu, Yang
    Yang, Feng
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2024, 73
  • [8] An object detection algorithm based on infrared-visible dual modal feature fusion
    Hou, Zhiqiang
    Yang, Chen
    Sun, Ying
    Ma, Sugang
    Yang, Xiaobao
    Fan, Jiulun
    INFRARED PHYSICS & TECHNOLOGY, 2024, 137
  • [9] Feature dynamic alignment and refinement for infrared-visible image fusion: Translation robust fusion
    Li, Huafeng
    Zhao, Junzhi
    Li, Jinxing
    Yu, Zhengtao
    Lu, Guangming
    INFORMATION FUSION, 2023, 95 : 26 - 41
  • [10] A Weighted Cross-Modal Feature Aggregation Network for Rumor Detection
    Li, Jia
    Hu, Zihan
    Yang, Zhenguo
    Lee, Lap-Kei
    Wang, Fu Lee
    ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PT VI, PAKDD 2024, 2024, 14650 : 42 - 53