CMIM: CROSS-MODAL INFORMATION MAXIMIZATION FOR MEDICAL IMAGING

被引:2
|
作者
Sylvain, Tristan [1 ,2 ]
Dutil, Francis [3 ]
Berthier, Tess [3 ]
Di Jorio, Lisa [3 ]
Luck, Margaux [1 ,2 ]
Hjelm, Devon [4 ]
Bengio, Yoshua [1 ,2 ]
机构
[1] Mila, Montreal, PQ, Canada
[2] Univ Montreal, Montreal, PQ, Canada
[3] Imagia Cybernet, Montreal, PQ, Canada
[4] Microsoft Res, Montreal, PQ, Canada
关键词
Deep learning; Medical Imaging; Multimodal data; Classification; Segmentation;
D O I
10.1109/ICASSP39728.2021.9414132
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
In hospitals, data are siloed to specific information systems that make the same information available under different modalities such as the different medical imaging exams the patient undergoes (CT scans, MRI, PET, Ultrasound, etc.) and their associated radiology reports. This offers unique opportunities to obtain and use at train-time those multiple views of the same information that might not always be available at test-time. In this paper, we propose an innovative framework that makes the most of available data by learning good representations of a multi-modal input that are resilient to modality dropping at test-time, using recent advances in mutual information maximization. By maximizing cross-modal information at train time, we are able to outperform several state-of-the-art baselines in two different settings, medical image classification, and segmentation. In particular, our method is shown to have a strong impact on the inference-time performance of weaker modalities.
引用
收藏
页码:1190 / 1194
页数:5
相关论文
共 50 条
  • [31] Cross-modal plasticity
    不详
    TRENDS IN COGNITIVE SCIENCES, 1997, 1 (07) : 251 - 251
  • [32] Cross-modal perception
    Zydlewska, Agnieszka
    Grabowska, Anna
    NEUROPSYCHIATRIA I NEUROPSYCHOLOGIA, 2011, 6 (02): : 60 - 70
  • [33] Enhancing Cross-Modal Medical Image Segmentation Through Compositionality
    Eijpe, Aniek
    Corbetta, Valentina
    Chupetlovska, Kalina
    Beets-Tan, Regina
    Silva, Wilson
    DEEP GENERATIVE MODELS, DGM4MICCAI 2024, 2025, 15224 : 43 - 53
  • [34] Cross-Modal Augmented Transformer for Automated Medical Report Generation
    Tang, Yuhao
    Yuan, Ye
    Tao, Fei
    Tang, Minghao
    IEEE JOURNAL OF TRANSLATIONAL ENGINEERING IN HEALTH AND MEDICINE, 2025, 13 : 33 - 48
  • [35] A cross-modal crowd counting method combining CNN and cross-modal transformer
    Zhang, Shihui
    Wang, Wei
    Zhao, Weibo
    Wang, Lei
    Li, Qunpeng
    IMAGE AND VISION COMPUTING, 2023, 129
  • [36] A semi-supervised cross-modal memory bank for cross-modal retrieval
    Huang, Yingying
    Hu, Bingliang
    Zhang, Yipeng
    Gao, Chi
    Wang, Quan
    NEUROCOMPUTING, 2024, 579
  • [37] Information Aggregation Semantic Adversarial Network for Cross-Modal Retrieval
    Wang, Hongfei
    Feng, Aimin
    Liu, Xuejun
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [38] KERNEL CROSS-MODAL FACTOR ANALYSIS FOR MULTIMODAL INFORMATION FUSION
    Wang, Yongjin
    Guan, Ling
    Venetsanopoulos, A. N.
    2011 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2011, : 2384 - 2387
  • [39] Cross-Modal Clustering With Deep Correlated Information Bottleneck Method
    Yan, Xiaoqiang
    Mao, Yiqiao
    Ye, Yangdong
    Yu, Hui
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (10) : 13508 - 13522
  • [40] Do early sensory cortices integrate cross-modal information?
    Kayser, Christoph
    Logothetis, Nikos K.
    BRAIN STRUCTURE & FUNCTION, 2007, 212 (02): : 121 - 132