CMIM: CROSS-MODAL INFORMATION MAXIMIZATION FOR MEDICAL IMAGING

被引:2
|
作者
Sylvain, Tristan [1 ,2 ]
Dutil, Francis [3 ]
Berthier, Tess [3 ]
Di Jorio, Lisa [3 ]
Luck, Margaux [1 ,2 ]
Hjelm, Devon [4 ]
Bengio, Yoshua [1 ,2 ]
机构
[1] Mila, Montreal, PQ, Canada
[2] Univ Montreal, Montreal, PQ, Canada
[3] Imagia Cybernet, Montreal, PQ, Canada
[4] Microsoft Res, Montreal, PQ, Canada
关键词
Deep learning; Medical Imaging; Multimodal data; Classification; Segmentation;
D O I
10.1109/ICASSP39728.2021.9414132
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
In hospitals, data are siloed to specific information systems that make the same information available under different modalities such as the different medical imaging exams the patient undergoes (CT scans, MRI, PET, Ultrasound, etc.) and their associated radiology reports. This offers unique opportunities to obtain and use at train-time those multiple views of the same information that might not always be available at test-time. In this paper, we propose an innovative framework that makes the most of available data by learning good representations of a multi-modal input that are resilient to modality dropping at test-time, using recent advances in mutual information maximization. By maximizing cross-modal information at train time, we are able to outperform several state-of-the-art baselines in two different settings, medical image classification, and segmentation. In particular, our method is shown to have a strong impact on the inference-time performance of weaker modalities.
引用
收藏
页码:1190 / 1194
页数:5
相关论文
共 50 条
  • [41] Integrating information theory and adversarial learning for cross-modal retrieval
    Chen, Wei
    Liu, Yu
    Bakker, Erwin M.
    Lew, Michael S.
    PATTERN RECOGNITION, 2021, 117
  • [42] Cross-Modal Center Loss for 3D Cross-Modal Retrieval
    Jing, Longlong
    Vahdani, Elahe
    Tan, Jiaxing
    Tian, Yingli
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 3141 - 3150
  • [43] Boon: A Neural Search Engine for Cross-Modal Information Retrieval
    Gong, Yan
    Cosma, Georgina
    PROCEEDINGS OF THE 1ST INTERNATIONAL WORKSHOP ON DEEP MULTIMODAL LEARNING FOR INFORMATION RETRIEVAL, MMIR 2023, 2023, : 29 - 37
  • [44] Similarity Retrieval and Medical Cross-Modal Attention Based Medical Report Generation
    Dong, Xinxin
    Pan, Haiwei
    Lan, Haiyan
    Zhang, Kejia
    Chen, Chunling
    WEB AND BIG DATA, APWEB-WAIM 2024, PT I, 2024, 14961 : 171 - 185
  • [45] Cross-Modal Information Aggregation and Distribution Method for Crowd Counting
    Chen, Yin
    Zhou, Yuhao
    Dong, Tianyang
    ADVANCES IN COMPUTER GRAPHICS, CGI 2023, PT IV, 2024, 14498 : 106 - 119
  • [46] Information Fusion via Deep Cross-Modal Factor Analysis
    Gao, Lei
    Guan, Ling
    2019 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2019,
  • [47] Multimodal Reaction: Information Modulation for Cross-Modal Representation Learning
    Zeng, Ying
    Mai, Sijie
    Yan, Wenjun
    Hu, Haifeng
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 2178 - 2191
  • [48] Do early sensory cortices integrate cross-modal information?
    Christoph Kayser
    Nikos K. Logothetis
    Brain Structure and Function, 2007, 212 : 121 - 132
  • [49] Cross-modal transfer of statistical information benefits from sleep
    Durrant, Simon J.
    Cairney, Scott A.
    Lewis, Penelope A.
    CORTEX, 2016, 78 : 85 - 99
  • [50] Topic correlation model for cross-modal multimedia information retrieval
    Qin, Zengchang
    Yu, Jing
    Cong, Yonghui
    Wan, Tao
    PATTERN ANALYSIS AND APPLICATIONS, 2016, 19 (04) : 1007 - 1022