Synthetic Augmentation with Large-Scale Unconditional Pre-training

被引:8
|
作者
Ye, Jiarong [1 ]
Ni, Haomiao [1 ]
Jin, Peng [1 ]
Huang, Sharon X. [1 ]
Xue, Yuan [2 ,3 ]
机构
[1] Penn State Univ, University Pk, PA 16802 USA
[2] Johns Hopkins Univ, Baltimore, MD 21218 USA
[3] Ohio State Univ, Columbus, OH 43210 USA
关键词
D O I
10.1007/978-3-031-43895-0_71
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep learning based medical image recognition systems often require a substantial amount of training data with expert annotations, which can be expensive and time-consuming to obtain. Recently, synthetic augmentation techniques have been proposed to mitigate the issue by generating realistic images conditioned on class labels. However, the effectiveness of these methods heavily depends on the representation capability of the trained generative model, which cannot be guaranteed without sufficient labeled training data. To further reduce the dependency on annotated data, we propose a synthetic augmentation method called HistoDiffusion, which can be pre-trained on large-scale unlabeled datasets and later applied to a small-scale labeled dataset for augmented training. In particular, we train a latent diffusion model (LDM) on diverse unlabeled datasets to learn common features and generate realistic images without conditional inputs. Then, we fine-tune the model with classifier guidance in latent space on an unseen labeled dataset so that the model can synthesize images of specific categories. Additionally, we adopt a selective mechanism to only add synthetic samples with high confidence of matching to target labels. We evaluate our proposed method by pre-training on three histopathology datasets and testing on a histopathology dataset of colorectal cancer (CRC) excluded from the pre-training datasets. With HistoDiffusion augmentation, the classification accuracy of a backbone classifier is remarkably improved by 6.4% using a small set of the original labels. Our code is available at https://github.com/karenyyy/HistoDiffAug.
引用
收藏
页码:754 / 764
页数:11
相关论文
共 50 条
  • [21] Robust feature learning for online discriminative tracking without large-scale pre-training
    Jun Zhang
    Bineng Zhong
    Pengfei Wang
    Cheng Wang
    Jixiang Du
    Frontiers of Computer Science, 2018, 12 : 1160 - 1172
  • [22] XCODE: Towards Cross-Language Code Representation with Large-Scale Pre-Training
    Lin, Zehao
    Li, Guodun
    Zhang, Jingfeng
    Deng, Yue
    Zeng, Xiangji
    Zhang, Yin
    Wan, Yao
    ACM TRANSACTIONS ON SOFTWARE ENGINEERING AND METHODOLOGY, 2022, 31 (03)
  • [23] WSPAlign: Word Alignment Pre-training via Large-Scale Weakly Supervised Span Prediction
    Wu, Qiyu
    Nagata, Masaaki
    Tsuruoka, Yoshimasa
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 11084 - 11099
  • [24] Editorial for Special Issue on Large-scale Pre-training: Data, Models, and Fine-tuning
    Wen, Ji-Rong
    Huang, Zi
    Zhang, Hanwang
    MACHINE INTELLIGENCE RESEARCH, 2023, 20 (02) : 145 - 146
  • [25] Learning meaningful representation of single-neuron morphology via large-scale pre-training
    Fan, Yimin
    Li, Yaxuan
    Zhong, Yunhua
    Hong, Liang
    Li, Lei
    Li, Yu
    BIOINFORMATICS, 2024, 40 : ii128 - ii136
  • [26] A Comparison between Pre-training and Large-scale Back-translation for Neural Machine Translation
    Huang, Dandan
    Wang, Kun
    Zhang, Yue
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 1718 - 1732
  • [27] Wukong: A 100 Million Large-scale Chinese Cross-modal Pre-training Benchmark
    Gu, Jiaxi
    Meng, Xiaojun
    Lu, Guansong
    Hou, Lu
    Niu, Minzhe
    Liang, Xiaodan
    Yao, Lewei
    Huang, Runhui
    Zhang, Wei
    Jiang, Xin
    Xu, Chunjing
    Xu, Hang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [28] WavLM: Large-Scale Self-Supervised Pre-Training for Full Stack Speech Processing
    Chen, Sanyuan
    Wang, Chengyi
    Chen, Zhengyang
    Wu, Yu
    Liu, Shujie
    Chen, Zhuo
    Li, Jinyu
    Kanda, Naoyuki
    Yoshioka, Takuya
    Xiao, Xiong
    Wu, Jian
    Zhou, Long
    Ren, Shuo
    Qian, Yanmin
    Qian, Yao
    Zeng, Michael
    Yu, Xiangzhan
    Wei, Furu
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2022, 16 (06) : 1505 - 1518
  • [29] Building a Strong Pre-Training Baseline for Universal 3D Large-Scale Perception
    Chen, Haoming
    Zhang, Zhizhong
    Qu, Yanyun
    Zhang, Ruixin
    Tan, Xin
    Xie, Yuan
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 19925 - 19935
  • [30] WenLan: Efficient Large-Scale Multi-Modal Pre-Training on Real World Data
    Song, Ruihua
    MMPT '21: PROCEEDINGS OF THE 2021 WORKSHOP ON MULTI-MODAL PRE-TRAINING FOR MULTIMEDIA UNDERSTANDING, 2021, : 3 - 3