Toward Unpaired Multi-modal Medical Image Segmentation via Learning Structured Semantic Consistency

被引:0
|
作者
Yang, Jie [1 ]
Zhu, Ye [1 ]
Wang, Chaoqun [1 ]
Li, Zhen [1 ]
Zhang, Ruimao [1 ]
机构
[1] Chinese Univ Hong Kong, Shenzhen Res Inst Big Data, Shenzhen, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Unpaired multi-modal learning; Structured semantic consistency learning; Medical image segmentation;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Integrating multi-modal data to promote medical image analysis has recently gained great attention. This paper presents a novel scheme to learn the mutual benefits of different modalities to achieve better segmentation results for unpaired multi-modal medical images. Our approach tackles two critical issues of this task from a practical perspective: (1) how to effectively learn the semantic consistencies of various modalities (e.g., CT and MRI), and (2) how to leverage the above consistencies to regularize the network learning while preserving its simplicity. To address (1), we leverage a carefully designed External Attention Module (EAM) to align semantic class representations and their correlations of different modalities. To solve (2), the proposed EAM is designed as an external plug-and-play one, which can be discarded once the model is optimized. We have demonstrated the effectiveness of the proposed method on two medical image segmentation scenarios: (1) cardiac structure segmentation, and (2) abdominal multi-organ segmentation. Extensive results show that the proposed method outperforms its counterparts by a wide margin.
引用
收藏
页码:1602 / 1622
页数:21
相关论文
共 50 条
  • [41] An overview of multi-modal medical image fusion
    Du, Jiao
    Li, Weisheng
    Lu, Ke
    Xiao, Bin
    NEUROCOMPUTING, 2016, 215 : 3 - 20
  • [42] Application of Multi-modal Fusion Attention Mechanism in Semantic Segmentation
    Liu, Yunlong
    Yoshie, Osamu
    Watanabe, Hiroshi
    COMPUTER VISION - ACCV 2022, PT VII, 2023, 13847 : 378 - 397
  • [43] Multi-modal Prototypes for Open-World Semantic Segmentation
    Yang, Yuhuan
    Ma, Chaofan
    Ju, Chen
    Zhang, Fei
    Yao, Jiangchao
    Zhang, Ya
    Wang, Yanfeng
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, 132 (12) : 6004 - 6020
  • [44] Semantic Segmentation of Defects in Infrastructures through Multi-modal Images
    Shahsavarani, Sara
    Lopez, Fernando
    Ibarra-Castanedo, Clemente
    Maldague, Xavier P., V
    THERMOSENSE: THERMAL INFRARED APPLICATIONS XLVI, 2024, 13047
  • [45] Ticino: A multi-modal remote sensing dataset for semantic segmentation
    Barbato, Mirko Paolo
    Piccoli, Flavio
    Napoletano, Paolo
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 249
  • [46] A framework for unsupervised segmentation of multi-modal medical images
    El-Baz, Ayman
    Farag, Aly
    Ali, Asem
    Gimel'farb, Georgy
    Casanova, Manuel
    COMPUTER VISION APPROACHES TO MEDICAL IMAGE ANALYSIS, 2006, 4241 : 120 - 131
  • [47] MULTI-MODAL MULTI-TASK LEARNING FOR SEMANTIC SEGMENTATION OF LAND COVER UNDER CLOUDY CONDITIONS
    Xu, Fang
    Shi, Yilei
    Yang, Wen
    Zhu, Xiaoxiang
    IGARSS 2023 - 2023 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM, 2023, : 6274 - 6277
  • [48] Comprehensive Multi-Modal Interactions for Referring Image Segmentation
    Jain, Kanishk
    Gandhi, Vineet
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), 2022, : 3427 - 3435
  • [49] Evaluation of Multi-modal Image Segmentation for Radiotherapy Planning
    Malladi, Sreeja
    Purohit, Sanket
    Brahme, Advait
    Scott, Julia A.
    IMAGING INFORMATICS FOR HEALTHCARE, RESEARCH, AND APPLICATIONS, MEDICAL IMAGING 2024, 2024, 12931
  • [50] Multi-task Learning of Semantic Segmentation and Height Estimation for Multi-modal Remote Sensing Images
    Mengyu WANG
    Zhiyuan YAN
    Yingchao FENG
    Wenhui DIAO
    Xian SUN
    Journal of Geodesy and Geoinformation Science, 2023, 6 (04) : 27 - 39