CMIM: CROSS-MODAL INFORMATION MAXIMIZATION FOR MEDICAL IMAGING

被引:2
|
作者
Sylvain, Tristan [1 ,2 ]
Dutil, Francis [3 ]
Berthier, Tess [3 ]
Di Jorio, Lisa [3 ]
Luck, Margaux [1 ,2 ]
Hjelm, Devon [4 ]
Bengio, Yoshua [1 ,2 ]
机构
[1] Mila, Montreal, PQ, Canada
[2] Univ Montreal, Montreal, PQ, Canada
[3] Imagia Cybernet, Montreal, PQ, Canada
[4] Microsoft Res, Montreal, PQ, Canada
关键词
Deep learning; Medical Imaging; Multimodal data; Classification; Segmentation;
D O I
10.1109/ICASSP39728.2021.9414132
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
In hospitals, data are siloed to specific information systems that make the same information available under different modalities such as the different medical imaging exams the patient undergoes (CT scans, MRI, PET, Ultrasound, etc.) and their associated radiology reports. This offers unique opportunities to obtain and use at train-time those multiple views of the same information that might not always be available at test-time. In this paper, we propose an innovative framework that makes the most of available data by learning good representations of a multi-modal input that are resilient to modality dropping at test-time, using recent advances in mutual information maximization. By maximizing cross-modal information at train time, we are able to outperform several state-of-the-art baselines in two different settings, medical image classification, and segmentation. In particular, our method is shown to have a strong impact on the inference-time performance of weaker modalities.
引用
收藏
页码:1190 / 1194
页数:5
相关论文
共 50 条
  • [1] Cross-modal image retrieval with deep mutual information maximization
    Gu, Chunbin
    Bu, Jiajun
    Zhou, Xixi
    Yao, Chengwei
    Ma, Dongfang
    Yu, Zhi
    Yan, Xifeng
    NEUROCOMPUTING, 2022, 496 : 166 - 177
  • [2] Cross-Modal Commentator: Automatic Machine Commenting Based on Cross-Modal Information
    Yang, Pengcheng
    Zhang, Zhihan
    Luo, Fuli
    Li, Lei
    Huang, Chengyang
    Sun, Xu
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 2680 - 2686
  • [3] Multimodal Mutual Information Maximization: A Novel Approach for Unsupervised Deep Cross-Modal Hashing
    Hoang, Tuan
    Do, Thanh-Toan
    Nguyen, Tam V.
    Cheung, Ngai-Man
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (09) : 6289 - 6302
  • [4] INFORMATION COMPLEXITY AND CROSS-MODAL FUNCTIONS
    FREIDES, D
    BRITISH JOURNAL OF PSYCHOLOGY, 1975, 66 (AUG) : 283 - 287
  • [5] Subspace learning by kernel dependence maximization for cross-modal retrieval
    Xu, Meixiang
    Zhu, Zhenfeng
    Zhao, Yao
    Sun, Fuming
    NEUROCOMPUTING, 2018, 309 : 94 - 105
  • [6] Deep medical cross-modal attention hashing
    Zhang, Yong
    Ou, Weihua
    Shi, Yufeng
    Deng, Jiaxin
    You, Xinge
    Wang, Anzhi
    WORLD WIDE WEB-INTERNET AND WEB INFORMATION SYSTEMS, 2022, 25 (04): : 1519 - 1536
  • [7] Deep medical cross-modal attention hashing
    Yong Zhang
    Weihua Ou
    Yufeng Shi
    Jiaxin Deng
    Xinge You
    Anzhi Wang
    World Wide Web, 2022, 25 : 1519 - 1536
  • [8] Cross-Modal Localization Through Mutual Information
    Alempijevic, Alen
    Kodagoda, Sarath
    Dissanayake, Gamini
    2009 IEEE-RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, 2009, : 5597 - 5602
  • [9] Cross-modal information integration in category learning
    Smith, J. David
    Johnston, Jennifer J. R.
    Musgrave, Robert D.
    Zakrzewski, Alexandria C.
    Boomer, Joseph
    Church, Barbara A.
    Ashby, F. Gregory
    ATTENTION PERCEPTION & PSYCHOPHYSICS, 2014, 76 (05) : 1473 - 1484
  • [10] Mechanism of Cross-modal Information Influencing Taste
    Liang, Pei
    Jiang, Jia-yu
    Liu, Qiang
    Zhang, Su-lin
    Yang, Hua-jing
    CURRENT MEDICAL SCIENCE, 2020, 40 (03) : 474 - 479