MSAFusionNet: Multiple Subspace Attention Based Deep Multi-modal Fusion Network

被引:4
|
作者
Zhang, Sen [1 ]
Zhang, Changzheng [1 ]
Wang, Lanjun [2 ]
Li, Cixing [1 ]
Tu, Dandan [1 ]
Luo, Rui [3 ]
Qi, Guojun [3 ]
Luo, Jiebo [4 ]
机构
[1] Huawei, Shenzhen, Peoples R China
[2] Huawei Canada, Markham, ON, Canada
[3] Futurewei, Bellevue, WA USA
[4] Univ Rochester, Rochester, NY 14627 USA
关键词
Deep learning; Multi-modal learning; Segmentation;
D O I
10.1007/978-3-030-32692-0_7
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
It is common for doctors to simultaneously consider multi-modal information in diagnosis. However, how to use multi-modal medical images effectively has not been fully studied in the field of deep learning within such a context. In this paper, we address the task of end-to-end segmentation based on multi-modal data and propose a novel deep learning framework, multiple subspace attention-based deep multi-modal fusion network (referred to as MSAFusionNet hereon-forth). More specifically, MSAFusionNet consists of three main components: (1) a multiple subspace attention model that contains inter-attention modules and generalized squeeze-and-excitation modules, (2) a multi-modal fusion network which leverages CNN-LSTM layers to integrate sequential multi-modal input images, and (3) a densely-dilated U-Net as the encoder-decoder backbone for image segmentation. Experiments on ISLES 2018 data set have shown that MSAFusionNet achieves the state-of-the-art segmentation accuracy.
引用
收藏
页码:54 / 62
页数:9
相关论文
共 50 条
  • [41] A multi-modal fusion YoLo network for traffic detection
    Zheng, Xinwang
    Zheng, Wenjie
    Xu, Chujie
    COMPUTATIONAL INTELLIGENCE, 2024, 40 (02)
  • [42] Electromagnetic signal feature fusion and recognition based on multi-modal deep learning
    Hou C.
    Zhang X.
    Chen X.
    International Journal of Performability Engineering, 2020, 16 (06): : 941 - 949
  • [43] Multi-modal Fusion Brain Tumor Detection Method Based on Deep Learning
    Yao Hong-ge
    Shen Xin-xia
    Li Yu
    Yu Jun
    Lei Song-ze
    ACTA PHOTONICA SINICA, 2019, 48 (07)
  • [44] Deep Learning Based Multi-Modal Fusion Architectures for Maritime Vessel Detection
    Farahnakian, Fahimeh
    Heikkonen, Jukka
    REMOTE SENSING, 2020, 12 (16)
  • [45] Cardiovascular disease detection based on deep learning and multi-modal data fusion
    Zhu, Jiayuan
    Liu, Hui
    Liu, Xiaowei
    Chen, Chao
    Shu, Minglei
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2025, 99
  • [46] DFAMNet: dual fusion attention multi-modal network for semantic segmentation on LiDAR point clouds
    Mingjie Li
    Gaihua Wang
    Minghao Zhu
    Chunzheng Li
    Hong Liu
    Xuran Pan
    Qian Long
    Applied Intelligence, 2024, 54 : 3169 - 3180
  • [47] TaGAT: Topology-Aware Graph Attention Network for Multi-modal Retinal Image Fusion
    Tian, Xin
    Anantrasirichai, Nantheera
    Nicholson, Lindsay
    Achim, Alin
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2024, PT I, 2024, 15001 : 775 - 784
  • [48] TAG-fusion: Two-stage attention guided multi-modal fusion network for semantic segmentation
    Zhang, Zhizhou
    Wang, Wenwu
    Zhu, Lei
    Tang, Zhibin
    DIGITAL SIGNAL PROCESSING, 2025, 156
  • [49] Deep-Learning-Based Multi-Modal Fusion for Fast MR Reconstruction
    Xiang, Lei
    Chen, Yong
    Chang, Weitang
    Zhan, Yiqiang
    Lin, Weili
    Wang, Qian
    Shen, Dinggang
    IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING, 2019, 66 (07) : 2105 - 2114
  • [50] DFAMNet: dual fusion attention multi-modal network for semantic segmentation on LiDAR point clouds
    Li, Mingjie
    Wang, Gaihua
    Zhu, Minghao
    Li, Chunzheng
    Liu, Hong
    Pan, Xuran
    Long, Qian
    APPLIED INTELLIGENCE, 2024, 54 (04) : 3169 - 3180