Self-supervised learning for chest computed tomography: training strategies and effect on downstream applications

被引:0
|
作者
Tariq, Amara [1 ]
Ramasamy, Gokul [1 ]
Patel, Bhavik [1 ,2 ,3 ]
Banerjee, Imon [1 ,2 ,3 ,4 ]
机构
[1] Mayo Clin Arizona, Arizona Adv AI Hub, Phoenix, AZ 85054 USA
[2] Mayo Clin Arizona, Dept Radiol, Phoenix, AZ USA
[3] Arizona State Univ, Sch Comp & Augmented Intelligence, Tempe, AZ USA
[4] Mayo Clin, Dept Artificial Intelligence & Informat, Scottsdale, AZ USA
关键词
biomedical imaging; computed tomography; image processing; self-supervised learning;
D O I
10.1117/1.JMI.11.6.064003
中图分类号
R8 [特种医学]; R445 [影像诊断学];
学科分类号
1002 ; 100207 ; 1009 ;
摘要
Purpose Self-supervised pre-training can reduce the amount of labeled training data needed by pre-learning fundamental visual characteristics of the medical imaging data. We investigate several self-supervised training strategies for chest computed tomography exams and their effects on downstream applications. Approach We benchmark five well-known self-supervision strategies (masked image region prediction, next slice prediction, rotation prediction, flip prediction, and denoising) on 15 M chest computed tomography (CT) slices collected from four sites of the Mayo Clinic enterprise, United States. These models were evaluated for two downstream tasks on public datasets: pulmonary embolism (PE) detection (classification) and lung nodule segmentation. Image embeddings generated by these models were also evaluated for prediction of patient age, race, and gender to study inherent biases in models' understanding of chest CT exams. Results The use of pre-training weights especially masked region prediction-based weights, improved performance, and reduced computational effort needed for downstream tasks compared with task-specific state-of-the-art (SOTA) models. Performance improvement for PE detection was observed for training dataset sizes as large as similar to 380 K with a maximum gain of 5% over SOTA. The segmentation model initialized with pre-training weights learned twice as fast as the randomly initialized model. While gender and age predictors built using self-supervised training weights showed no performance improvement over randomly initialized predictors, the race predictor experienced a 10% performance boost when using self-supervised training weights. Conclusion We released self-supervised models and weights under an open-source academic license. These models can then be fine-tuned with limited task-specific annotated data for a variety of downstream imaging tasks, thus accelerating research in biomedical imaging informatics.
引用
收藏
页数:19
相关论文
共 50 条
  • [41] Dense Contrastive Learning for Self-Supervised Visual Pre-Training
    Wang, Xinlong
    Zhang, Rufeng
    Shen, Chunhua
    Kong, Tao
    Li, Lei
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 3023 - 3032
  • [42] Self-supervised Learning for MRI Reconstruction with a Parallel Network Training Framework
    Hu, Chen
    Li, Cheng
    Wang, Haifeng
    Liu, Qiegen
    Zheng, Hairong
    Wang, Shanshan
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2021, PT VI, 2021, 12906 : 382 - 391
  • [43] TRAINING SET CLEANSING OF BACKDOOR POISONING BY SELF-SUPERVISED REPRESENTATION LEARNING
    Wang, Hang
    Karami, Sahar
    Dia, Ousmane
    Ritter, Hippolyt
    Emamjomeh-Zadeh, Ehsan
    Chen, Jiahui
    Xiang, Zhen
    Miller, David J.
    Kesidis, George
    arXiv, 2022,
  • [44] Biologically Plausible Training Mechanisms for Self-Supervised Learning in Deep Networks
    Tang, Mufeng
    Yang, Yibo
    Amit, Yali
    FRONTIERS IN COMPUTATIONAL NEUROSCIENCE, 2022, 16
  • [45] METRICBERT: TEXT REPRESENTATION LEARNING VIA SELF-SUPERVISED TRIPLET TRAINING
    Malkiel, Itzik
    Ginzburg, Dvir
    Barkan, Oren
    Caciularu, Avi
    Weill, Yoni
    Koenigstein, Noam
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 8142 - 8146
  • [46] Training Set Cleansing of Backdoor Poisoning by Self-Supervised Representation Learning
    Wang, Hang
    Karami, Sahar
    Dia, Ousmane
    Ritter, Hippolyt
    Emamjomeh-Zadeh, Ehsan
    Chen, Jiahui
    Xiang, Zhen
    Miller, David J.
    Kesidis, George
    ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings, 2023,
  • [47] Self-supervised denoising of grating-based phase-contrast computed tomography
    Wirtensohn, Sami
    Schmid, Clemens
    Berthe, Daniel
    John, Dominik
    Heck, Lisa
    Taphorn, Kirsten
    Flenner, Silja
    Herzen, Julia
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [48] Dual-Domain Self-Supervised Deep Learning with Graph Convolution for Low-Dose Computed Tomography Reconstruction
    Yang, Feng
    Zhao, Feixiang
    Liu, Yanhua
    Liu, Min
    Liu, Mingzhe
    JOURNAL OF IMAGING INFORMATICS IN MEDICINE, 2025,
  • [49] Semi- and Self-Supervised Metric Learning for Remote Sensing Applications
    Hernandez-Sequeira, Itza
    Fernandez-Beltran, Ruben
    Pla, Filiberto
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2024, 21 : 1 - 5
  • [50] Self-supervised Pre-training and Semi-supervised Learning for Extractive Dialog Summarization
    Zhuang, Yingying
    Song, Jiecheng
    Sadagopan, Narayanan
    Beniwal, Anurag
    COMPANION OF THE WORLD WIDE WEB CONFERENCE, WWW 2023, 2023, : 1069 - 1076