MSAFusionNet: Multiple Subspace Attention Based Deep Multi-modal Fusion Network

被引:4
|
作者
Zhang, Sen [1 ]
Zhang, Changzheng [1 ]
Wang, Lanjun [2 ]
Li, Cixing [1 ]
Tu, Dandan [1 ]
Luo, Rui [3 ]
Qi, Guojun [3 ]
Luo, Jiebo [4 ]
机构
[1] Huawei, Shenzhen, Peoples R China
[2] Huawei Canada, Markham, ON, Canada
[3] Futurewei, Bellevue, WA USA
[4] Univ Rochester, Rochester, NY 14627 USA
关键词
Deep learning; Multi-modal learning; Segmentation;
D O I
10.1007/978-3-030-32692-0_7
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
It is common for doctors to simultaneously consider multi-modal information in diagnosis. However, how to use multi-modal medical images effectively has not been fully studied in the field of deep learning within such a context. In this paper, we address the task of end-to-end segmentation based on multi-modal data and propose a novel deep learning framework, multiple subspace attention-based deep multi-modal fusion network (referred to as MSAFusionNet hereon-forth). More specifically, MSAFusionNet consists of three main components: (1) a multiple subspace attention model that contains inter-attention modules and generalized squeeze-and-excitation modules, (2) a multi-modal fusion network which leverages CNN-LSTM layers to integrate sequential multi-modal input images, and (3) a densely-dilated U-Net as the encoder-decoder backbone for image segmentation. Experiments on ISLES 2018 data set have shown that MSAFusionNet achieves the state-of-the-art segmentation accuracy.
引用
收藏
页码:54 / 62
页数:9
相关论文
共 50 条
  • [31] Multi-modal Fusion
    Liu, Huaping
    Hussain, Amir
    Wang, Shuliang
    INFORMATION SCIENCES, 2018, 432 : 462 - 462
  • [32] AMM-FuseNet: Attention-Based Multi-Modal Image Fusion Network for Land Cover Mapping
    Ma, Wanli
    Karaku, Oktay
    Rosin, Paul L.
    REMOTE SENSING, 2022, 14 (18)
  • [33] Attention-based Fusion Network for Breast Cancer Segmentation and Classification Using Multi-modal Ultrasound Images
    Cho, Yoonjae
    Misra, Sampa
    Managuli, Ravi
    Barr, Richard G.
    Lee, Jeongmin
    Kim, Chulhong
    ULTRASOUND IN MEDICINE AND BIOLOGY, 2025, 51 (03): : 568 - 577
  • [34] Application of Multi-modal Fusion Attention Mechanism in Semantic Segmentation
    Liu, Yunlong
    Yoshie, Osamu
    Watanabe, Hiroshi
    COMPUTER VISION - ACCV 2022, PT VII, 2023, 13847 : 378 - 397
  • [35] Multi-modal Conditional Attention Fusion for Dimensional Emotion Prediction
    Chen, Shizhe
    Jin, Qin
    MM'16: PROCEEDINGS OF THE 2016 ACM MULTIMEDIA CONFERENCE, 2016, : 571 - 575
  • [36] Multi-Modal Subspace Fusion via Cauchy Multi-Set Canonical Correlations
    Zhu, Yanmin
    Peng, Tianhao
    Su, Shuzhi
    Li, Changpeng
    IEEE ACCESS, 2020, 8 : 115228 - 115239
  • [37] Deep multi-modal fusion network with gated unit for breast cancer survival prediction
    Yuan, Han
    Xu, Hongzhen
    COMPUTER METHODS IN BIOMECHANICS AND BIOMEDICAL ENGINEERING, 2024, 27 (07) : 883 - 896
  • [38] Deep Multi-Modal Network Based Automated Depression Severity Estimation
    Uddin, Md Azher
    Joolee, Joolekha Bibi
    Sohn, Kyung-Ah
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2023, 14 (03) : 2153 - 2167
  • [39] AGGN: Attention-based glioma grading network with multi-scale feature extraction and multi-modal information fusion
    Wu, Peishu
    Wang, Zidong
    Zheng, Baixun
    Li, Han
    Alsaadi, Fuad E.
    Zeng, Nianyin
    COMPUTERS IN BIOLOGY AND MEDICINE, 2023, 152
  • [40] Cross-modal attention network for retinal disease classification based on multi-modal images
    Liu, Zirong
    Hu, Yan
    Qiu, Zhongxi
    Niu, Yanyan
    Zhou, Dan
    Li, Xiaoling
    Shen, Junyong
    Jiang, Hongyang
    Li, Heng
    Liu, Jiang
    BIOMEDICAL OPTICS EXPRESS, 2024, 15 (06): : 3699 - 3714