MATR: Multimodal Medical Image Fusion via Multiscale Adaptive Transformer

被引:0
|
作者
Tang, Wei [1 ]
He, Fazhi [1 ]
Liu, Yu [2 ]
Duan, Yansong [3 ]
机构
[1] Wuhan University, School of Computer Science, Wuhan,430072, China
[2] Hefei University of Technology, Department of Biomedical Engineering, Hefei,230009, China
[3] Wuhan University, School of Remote Sensing and Information Engineering, Wuhan,430079, China
关键词
Convolution - Deep learning - Diagnosis - Image fusion - Medical imaging - Network architecture - Particle beams - Quality control - Semantics;
D O I
暂无
中图分类号
学科分类号
摘要
Owing to the limitations of imaging sensors, it is challenging to obtain a medical image that simultaneously contains functional metabolic information and structural tissue details. Multimodal medical image fusion, an effective way to merge the complementary information in different modalities, has become a significant technique to facilitate clinical diagnosis and surgical navigation. With powerful feature representation ability, deep learning (DL)-based methods have improved such fusion results but still have not achieved satisfactory performance. Specifically, existing DL-based methods generally depend on convolutional operations, which can well extract local patterns but have limited capability in preserving global context information. To compensate for this defect and achieve accurate fusion, we propose a novel unsupervised method to fuse multimodal medical images via a multiscale adaptive Transformer termed MATR. In the proposed method, instead of directly employing vanilla convolution, we introduce an adaptive convolution for adaptively modulating the convolutional kernel based on the global complementary context. To further model long-range dependencies, an adaptive Transformer is employed to enhance the global semantic extraction capability. Our network architecture is designed in a multiscale fashion so that useful multimodal information can be adequately acquired from the perspective of different scales. Moreover, an objective function composed of a structural loss and a region mutual information loss is devised to construct constraints for information preservation at both the structural-level and the feature-level. Extensive experiments on a mainstream database demonstrate that the proposed method outperforms other representative and state-of-the-art methods in terms of both visual quality and quantitative evaluation. We also extend the proposed method to address other biomedical image fusion issues, and the pleasing fusion results illustrate that MATR has good generalization capability. The code of the proposed method is available at https://github.com/tthinking/MATR. © 1992-2012 IEEE.
引用
收藏
页码:5134 / 5149
相关论文
共 50 条
  • [1] MATR: Multimodal Medical Image Fusion via Multiscale Adaptive Transformer
    Tang, Wei
    He, Fazhi
    Liu, Yu
    Duan, Yansong
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 : 5134 - 5149
  • [2] MRSCFusion: Joint Residual Swin Transformer and Multiscale CNN for Unsupervised Multimodal Medical Image Fusion
    Xie, Xinyu
    Zhang, Xiaozhi
    Ye, Shengcheng
    Xiong, Dongping
    Ouyang, Lijun
    Yang, Bin
    Zhou, Hong
    Wan, Yaping
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2023, 72
  • [3] Multimodal Medical Image Fusion by Adaptive Manifold Filter
    Geng, Peng
    Liu, Shuaiqi
    Zhuang, Shanna
    COMPUTATIONAL AND MATHEMATICAL METHODS IN MEDICINE, 2015, 2015
  • [4] Multimodal Image Fusion via Self-Supervised Transformer
    Zhang, Jing
    Liu, Yu
    Liu, Aiping
    Xie, Qingguo
    Ward, Rabab
    Wang, Z. Jane
    Chen, Xun
    IEEE SENSORS JOURNAL, 2023, 23 (09) : 9796 - 9807
  • [5] SMTF: Sparse transformer with multiscale contextual fusion for medical image segmentation
    Zhang, Xichu
    Zhang, Xiaozhi
    Ouyang, Lijun
    Qin, Chuanbo
    Xiao, Lin
    Xiong, Dongping
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2024, 87
  • [6] Gradient Compass-Based Adaptive Multimodal Medical Image Fusion
    Irshad, Muhammad Touseef
    Rehman, Hafeez Ur
    IEEE ACCESS, 2021, 9 : 22662 - 22670
  • [7] A multibranch and multiscale neural network based on semantic perception for multimodal medical image fusion
    Lin, Cong
    Chen, Yinjie
    Feng, Siling
    Huang, Mengxing
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [8] Multimodal Image Fusion Method Based on Multiscale Image Matting
    Maqsood, Sarmad
    Damasevicius, Robertas
    Silka, Jakub
    Wozniak, Marcin
    ARTIFICIAL INTELLIGENCE AND SOFT COMPUTING (ICAISC 2021), PT II, 2021, 12855 : 57 - 68
  • [9] Automatic multimodal medical image fusion
    Zhang, ZF
    Yao, J
    Bajwa, S
    Gudas, T
    SMCIA/03: PROCEEDINGS OF THE 2003 IEEE INTERNATIONAL WORKSHOP ON SOFT COMPUTING IN INDUSTRIAL APPLICATIONS, 2003, : 161 - 166
  • [10] A review on multimodal medical image fusion
    Reddy, G. R. Byra
    Kumar, H. Prasanna
    INTERNATIONAL JOURNAL OF BIOMEDICAL ENGINEERING AND TECHNOLOGY, 2020, 34 (02) : 119 - 132