Synthetic Augmentation with Large-Scale Unconditional Pre-training

被引:8
|
作者
Ye, Jiarong [1 ]
Ni, Haomiao [1 ]
Jin, Peng [1 ]
Huang, Sharon X. [1 ]
Xue, Yuan [2 ,3 ]
机构
[1] Penn State Univ, University Pk, PA 16802 USA
[2] Johns Hopkins Univ, Baltimore, MD 21218 USA
[3] Ohio State Univ, Columbus, OH 43210 USA
关键词
D O I
10.1007/978-3-031-43895-0_71
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep learning based medical image recognition systems often require a substantial amount of training data with expert annotations, which can be expensive and time-consuming to obtain. Recently, synthetic augmentation techniques have been proposed to mitigate the issue by generating realistic images conditioned on class labels. However, the effectiveness of these methods heavily depends on the representation capability of the trained generative model, which cannot be guaranteed without sufficient labeled training data. To further reduce the dependency on annotated data, we propose a synthetic augmentation method called HistoDiffusion, which can be pre-trained on large-scale unlabeled datasets and later applied to a small-scale labeled dataset for augmented training. In particular, we train a latent diffusion model (LDM) on diverse unlabeled datasets to learn common features and generate realistic images without conditional inputs. Then, we fine-tune the model with classifier guidance in latent space on an unseen labeled dataset so that the model can synthesize images of specific categories. Additionally, we adopt a selective mechanism to only add synthetic samples with high confidence of matching to target labels. We evaluate our proposed method by pre-training on three histopathology datasets and testing on a histopathology dataset of colorectal cancer (CRC) excluded from the pre-training datasets. With HistoDiffusion augmentation, the classification accuracy of a backbone classifier is remarkably improved by 6.4% using a small set of the original labels. Our code is available at https://github.com/karenyyy/HistoDiffAug.
引用
收藏
页码:754 / 764
页数:11
相关论文
共 50 条
  • [1] Pre-training on Large-Scale Heterogeneous Graph
    Jiang, Xunqiang
    Jia, Tianrui
    Fang, Yuan
    Shi, Chuan
    Lin, Zhe
    Wang, Hui
    KDD '21: PROCEEDINGS OF THE 27TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2021, : 756 - 766
  • [2] Automating Code Review Activities by Large-Scale Pre-training
    Li, Zhiyu
    Lu, Shuai
    Guo, Daya
    Duan, Nan
    Jannu, Shailesh
    Jenks, Grant
    Majumder, Deep
    Green, Jared
    Svyatkovskiy, Alexey
    Fu, Shengyu
    Sundaresan, Neel
    PROCEEDINGS OF THE 30TH ACM JOINT MEETING EUROPEAN SOFTWARE ENGINEERING CONFERENCE AND SYMPOSIUM ON THE FOUNDATIONS OF SOFTWARE ENGINEERING, ESEC/FSE 2022, 2022, : 1035 - 1047
  • [3] An Optimized Method for Large-Scale Pre-Training in Symbolic Music
    Liu, Shike
    Xu, Hongguang
    Xu, Ke
    Proceedings of the International Conference on Anti-Counterfeiting, Security and Identification, ASID, 2022, 2022-December : 105 - 109
  • [4] PreDet: Large-scale weakly supervised pre-training for detection
    Ramanathan, Vignesh
    Wang, Rui
    Mahajan, Dhruv
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 2845 - 2855
  • [5] MusicBERT: Symbolic Music Understanding with Large-Scale Pre-Training
    Zeng, Mingliang
    Tan, Xu
    Wang, Rui
    Ju, Zeqian
    Qin, Tao
    Liu, Tie-Yan
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 791 - 800
  • [6] MuralDiff: Diffusion for Ancient Murals Restoration on Large-Scale Pre-Training
    Xu, Zishan
    Zhang, Xiaofeng
    Chen, Wei
    Liu, Jueting
    Xu, Tingting
    Wang, Zehua
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2024, 8 (03): : 2169 - 2181
  • [7] DIALOGPT : Large-Scale Generative Pre-training for Conversational Response Generation
    Zhang, Yizhe
    Sun, Siqi
    Galley, Michel
    Chen, Yen-Chun
    Brockett, Chris
    Gao, Xiang
    Gao, Jianfeng
    Liu, Jingjing
    Dolan, Bill
    58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020): SYSTEM DEMONSTRATIONS, 2020, : 270 - 278
  • [8] BigDetection: A Large-scale Benchmark for Improved Object Detector Pre-training
    Cai, Likun
    Zhang, Zhi
    Zhu, Yi
    Zhang, Li
    Li, Mu
    Xue, Xiangyang
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2022, 2022, : 4776 - 4786
  • [9] Large-scale weakly-supervised pre-training for video action recognition
    Ghadiyaram, Deepti
    Du Tran
    Mahajan, Dhruv
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 12038 - 12047
  • [10] Large-Scale Unsupervised Audio Pre-Training for Video-to-Speech Synthesis
    Kefalas, Triantafyllos
    Panagakis, Yannis
    Pantic, Maja
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 2255 - 2268