MSAFusionNet: Multiple Subspace Attention Based Deep Multi-modal Fusion Network

被引:4
|
作者
Zhang, Sen [1 ]
Zhang, Changzheng [1 ]
Wang, Lanjun [2 ]
Li, Cixing [1 ]
Tu, Dandan [1 ]
Luo, Rui [3 ]
Qi, Guojun [3 ]
Luo, Jiebo [4 ]
机构
[1] Huawei, Shenzhen, Peoples R China
[2] Huawei Canada, Markham, ON, Canada
[3] Futurewei, Bellevue, WA USA
[4] Univ Rochester, Rochester, NY 14627 USA
关键词
Deep learning; Multi-modal learning; Segmentation;
D O I
10.1007/978-3-030-32692-0_7
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
It is common for doctors to simultaneously consider multi-modal information in diagnosis. However, how to use multi-modal medical images effectively has not been fully studied in the field of deep learning within such a context. In this paper, we address the task of end-to-end segmentation based on multi-modal data and propose a novel deep learning framework, multiple subspace attention-based deep multi-modal fusion network (referred to as MSAFusionNet hereon-forth). More specifically, MSAFusionNet consists of three main components: (1) a multiple subspace attention model that contains inter-attention modules and generalized squeeze-and-excitation modules, (2) a multi-modal fusion network which leverages CNN-LSTM layers to integrate sequential multi-modal input images, and (3) a densely-dilated U-Net as the encoder-decoder backbone for image segmentation. Experiments on ISLES 2018 data set have shown that MSAFusionNet achieves the state-of-the-art segmentation accuracy.
引用
收藏
页码:54 / 62
页数:9
相关论文
共 50 条
  • [1] Dual-Attention Deep Fusion Network for Multi-modal Medical Image Segmentation
    Zheng, Shenhai
    Ye, Xin
    Tan, Jiaxin
    Yang, Yifei
    Li, Laquan
    FOURTEENTH INTERNATIONAL CONFERENCE ON GRAPHICS AND IMAGE PROCESSING, ICGIP 2022, 2022, 12705
  • [2] Attention-Based Multi-Modal Fusion Network for Semantic Scene Completion
    Li, Siqi
    Zou, Changqing
    Li, Yipeng
    Zhao, Xibin
    Gao, Yue
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 11402 - 11409
  • [3] Robust Deep Multi-modal Learning Based on Gated Information Fusion Network
    Kim, Jaekyum
    Koh, Junho
    Kim, Yecheol
    Choi, Jaehyung
    Hwang, Youngbae
    Choi, Jun Won
    COMPUTER VISION - ACCV 2018, PT IV, 2019, 11364 : 90 - 106
  • [4] ISAFusionNet: Involution and soft attention based deep multi-modal fusion network for multi-label skin lesion classification
    Mohammed, Hussein M. A.
    Omeroglu, Asli Nur
    Oral, Emin Argun
    Ozbek, I. Yucel
    COMPUTERS & ELECTRICAL ENGINEERING, 2025, 122
  • [5] Memory based fusion for multi-modal deep learning
    Priyasad, Darshana
    Fernando, Tharindu
    Denman, Simon
    Sridharan, Sridha
    Fookes, Clinton
    INFORMATION FUSION, 2021, 67 : 136 - 146
  • [6] Multi-Modal Fusion Sign Language Recognition Based on Residual Network and Attention Mechanism
    Chu Chaoqin
    Xiao Qinkun
    Zhang Yinhuan
    Xing, Liu
    INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2022, 36 (12)
  • [7] A co-attention based multi-modal fusion network for review helpfulness prediction
    Ren, Gang
    Diao, Lei
    Guo, Fanjia
    Hong, Taeho
    INFORMATION PROCESSING & MANAGEMENT, 2024, 61 (01)
  • [8] A Tri-Attention fusion guided multi-modal segmentation network
    Zhou, Tongxue
    Ruan, Su
    Vera, Pierre
    Canu, Stephane
    PATTERN RECOGNITION, 2022, 124
  • [9] Attention-based multi-modal fusion sarcasm detection
    Liu, Jing
    Tian, Shengwei
    Yu, Long
    Long, Jun
    Zhou, Tiejun
    Wang, Bo
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2023, 44 (02) : 2097 - 2108
  • [10] Multi-modal Perception Fusion Method Based on Cross Attention
    Zhang B.-L.
    Pan Z.-H.
    Jiang J.-Z.
    Zhang C.-B.
    Wang Y.-X.
    Yang C.-L.
    Zhongguo Gonglu Xuebao/China Journal of Highway and Transport, 2024, 37 (03): : 181 - 193