Self-supervised learning for chest computed tomography: training strategies and effect on downstream applications

被引:0
|
作者
Tariq, Amara [1 ]
Ramasamy, Gokul [1 ]
Patel, Bhavik [1 ,2 ,3 ]
Banerjee, Imon [1 ,2 ,3 ,4 ]
机构
[1] Mayo Clin Arizona, Arizona Adv AI Hub, Phoenix, AZ 85054 USA
[2] Mayo Clin Arizona, Dept Radiol, Phoenix, AZ USA
[3] Arizona State Univ, Sch Comp & Augmented Intelligence, Tempe, AZ USA
[4] Mayo Clin, Dept Artificial Intelligence & Informat, Scottsdale, AZ USA
关键词
biomedical imaging; computed tomography; image processing; self-supervised learning;
D O I
10.1117/1.JMI.11.6.064003
中图分类号
R8 [特种医学]; R445 [影像诊断学];
学科分类号
1002 ; 100207 ; 1009 ;
摘要
Purpose Self-supervised pre-training can reduce the amount of labeled training data needed by pre-learning fundamental visual characteristics of the medical imaging data. We investigate several self-supervised training strategies for chest computed tomography exams and their effects on downstream applications. Approach We benchmark five well-known self-supervision strategies (masked image region prediction, next slice prediction, rotation prediction, flip prediction, and denoising) on 15 M chest computed tomography (CT) slices collected from four sites of the Mayo Clinic enterprise, United States. These models were evaluated for two downstream tasks on public datasets: pulmonary embolism (PE) detection (classification) and lung nodule segmentation. Image embeddings generated by these models were also evaluated for prediction of patient age, race, and gender to study inherent biases in models' understanding of chest CT exams. Results The use of pre-training weights especially masked region prediction-based weights, improved performance, and reduced computational effort needed for downstream tasks compared with task-specific state-of-the-art (SOTA) models. Performance improvement for PE detection was observed for training dataset sizes as large as similar to 380 K with a maximum gain of 5% over SOTA. The segmentation model initialized with pre-training weights learned twice as fast as the randomly initialized model. While gender and age predictors built using self-supervised training weights showed no performance improvement over randomly initialized predictors, the race predictor experienced a 10% performance boost when using self-supervised training weights. Conclusion We released self-supervised models and weights under an open-source academic license. These models can then be fine-tuned with limited task-specific annotated data for a variety of downstream imaging tasks, thus accelerating research in biomedical imaging informatics.
引用
收藏
页数:19
相关论文
共 50 条
  • [1] Unsupervised and Self-supervised Learning in Low-Dose Computed Tomography Denoising: Insights from Training Strategies
    Zhao, Feixiang
    Liu, Mingzhe
    Xiang, Mingrong
    Li, Dongfen
    Jiang, Xin
    Jin, Xiance
    Lin, Cai
    Wang, Ruili
    JOURNAL OF IMAGING INFORMATICS IN MEDICINE, 2025, 38 (02): : 902 - 930
  • [2] Accelerating Self-Supervised Learning via Efficient Training Strategies
    Kocyigit, Mustafa Taha
    Hospedales, Timothy M.
    Bilen, Hakan
    2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 5643 - 5653
  • [3] Self-Adaptive Training: Bridging Supervised and Self-Supervised Learning
    Huang, Lang
    Zhang, Chao
    Zhang, Hongyang
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (03) : 1362 - 1377
  • [4] Self-supervised Physics-based Denoising for Computed Tomography
    Zainulina, Elvira
    Chernyavskiy, Alexey
    Dylov, Dmitry V.
    arXiv, 2022,
  • [5] Self-supervised Contrastive Learning for Predicting Game Strategies
    Lee, Young Jae
    Baek, Insung
    Jo, Uk
    Kim, Jaehoon
    Bae, Jinsoo
    Jeong, Keewon
    Kim, Seoung Bum
    INTELLIGENT SYSTEMS AND APPLICATIONS, VOL 1, 2023, 542 : 136 - 147
  • [6] Applications of Self-Supervised Learning to Biomedical Signals: A Survey
    Del Pup, Federico
    Atzori, Manfredo
    IEEE ACCESS, 2023, 11 : 144180 - 144203
  • [7] Self-supervised Learning of LiDAR Odometry for Robotic Applications
    Nubert, Julian
    Khattak, Shehryar
    Hutter, Marco
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 9601 - 9607
  • [8] Masked cross-domain self-supervised deep learning framework for photoacoustic computed tomography reconstruction
    Lan, Hengrong
    Huang, Lijie
    Wei, Xingyue
    Li, Zhiqiang
    Lv, Jing
    Ma, Cheng
    Nie, Liming
    Luo, Jianwen
    NEURAL NETWORKS, 2024, 179
  • [9] SELF-SUPERVISED ADVERSARIAL TRAINING
    Chen, Kejiang
    Chen, Yuefeng
    Zhou, Hang
    Mao, Xiaofeng
    Li, Yuhong
    He, Yuan
    Xue, Hui
    Zhang, Weiming
    Yu, Nenghai
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 2218 - 2222
  • [10] Pareto Self-Supervised Training for Few-Shot Learning
    Chen, Zhengyu
    Ge, Jixie
    Zhan, Heshen
    Huang, Siteng
    Wang, Donglin
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 13658 - 13667