Research on Dual-adversarial MR Image Fusion Network Using Pre-trained Model for Feature Extraction

被引:0
|
作者
Liu H. [1 ,2 ]
Li S.-S. [1 ,2 ]
Gao S.-S. [1 ,2 ]
Deng K. [3 ]
Xu G. [4 ]
Zhang C.-M. [2 ,5 ]
机构
[1] School of Computer Science and Technology, Shandong University of Finance and Economics, Jinan
[2] Shandong Key Laboratory of Digital Media Technology, Jinan
[3] The First Affiliated Hospital of Shandong First Medical University, Jinan
[4] College of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou
[5] School of Software, Shandong University, Jinan
来源
Ruan Jian Xue Bao/Journal of Software | 2023年 / 34卷 / 05期
关键词
adversarial learning; dual-adversarial network; image fusion; multi-modal medical image; pre-trained model;
D O I
10.13328/j.cnki.jos.006772
中图分类号
学科分类号
摘要
With the popularization of multimodal medical images in clinical diagnosis and treatment, fusion technology based on spatiotemporal correlation characteristics has been developed rapidly. The fused medical images not only retain the unique features of source images with various modalities but also strengthen the complementary information, which can facilitate image reading. At present, most methods perform feature extraction and feature fusion by manually defining constraints, which can easily lead to the loss of useful information and unclear details in the fused images. In light of this, a dual-adversarial fusion network using a pre-trained model for feature extraction is proposed in this study to fuse MR-T1/MR-T2 images. The network consists of a feature extraction module, a feature fusion module, and two discriminator network modules. Due to the small scale of the registered multimodal medical image dataset, the feature extraction network cannot be fully trained. In addition, as the pre-trained model has powerful data representation ability, a pre-trained convolutional neural network model is embedded into the feature extraction module to generate the feature map. Then, the feature fusion network fuses the deep features and outputs fused images. Through accurate classification of the source and fused images, the two discriminator networks establish adversarial relations with the feature fusion network separately and eventually encourage it to learn the optimal fusion parameters. The experimental results illustrate the effectiveness of pre-trained technology in this method. Compared with six existing typical fusion methods, the proposed method can generate the fused results of optimal performance in visual effects and quantitative metrics. © 2023 Chinese Academy of Sciences. All rights reserved.
引用
收藏
页码:2134 / 2151
页数:17
相关论文
共 47 条
  • [1] Liu H, Xu J, Wu Y, Guo Q, Ibragimov B, Xing L., Learning deconvolutional deep neural network for high resolution medical image reconstruction, Information Sciences, 468, pp. 142-154, (2018)
  • [2] Liu H, Wang HO, Wu Y, Xing L., Superpixel region merging based on deep network for medical image segmentation, ACM Trans. on Intelligent Systems and Technology, 11, 4, pp. 1-22, (2020)
  • [3] Yu X, Liu H, Wu Y, Zhang CM., Intrinsic self-representation for multi-view subspace clustering, Scienta Sinica Informations, 51, 10, pp. 1625-1639, (2021)
  • [4] Hermessi H, Mourali O, Zagrouba E., Multimodal medical image fusion review: Theoretical background and recent advances, Signal Processing, 183, (2021)
  • [5] Zhang H, Xu H, Tian X, Jiang JJ, Ma JY., Image fusion meets deep learning: A survey and perspective, Information Fusion, 76, pp. 323-336, (2021)
  • [6] Li Y, Zhao JL, Lv ZH, Li JH., Medical image fusion method by deep learning, Int’l Journal of Cognitive Computing in Engineering, 2, pp. 21-29, (2021)
  • [7] Prabhakar KR, Srikar VS, Babu RV., DeepFuse: A deep unsupervised approach for exposure fusion with extreme exposure image pairs, Proc. of the 2017 IEEE Int’l Conf. on Computer Vision, pp. 4714-4722, (2017)
  • [8] Li H, Wu XJ., DenseFuse: A fusion approach to infrared and visible images, IEEE Trans. on Image Processing, 28, 5, pp. 2614-2623, (2019)
  • [9] Han X, Zhang ZY, Ding N, Gu YX, Liu X, Huo YQ, Qiu JZ, Yao Y, Zhang A, Zhang L, Han WT, Huang ML, Jin Q, Lan YY, Liu Y, Liu ZY, Lu ZW, Qiu XP, Song RH, Tang J, Wen JR, Yuan JH, Zhao WX, Zhu J., Pre-trained models: Past, present and future, AI Open, 2, pp. 225-250, (2021)
  • [10] Iglovikov V, Shvets A., TernausNet: U-Net with VGG11 encoder pre-trained on ImageNet for image segmentation, (2018)