Synthetic Augmentation with Large-Scale Unconditional Pre-training

被引:8
|
作者
Ye, Jiarong [1 ]
Ni, Haomiao [1 ]
Jin, Peng [1 ]
Huang, Sharon X. [1 ]
Xue, Yuan [2 ,3 ]
机构
[1] Penn State Univ, University Pk, PA 16802 USA
[2] Johns Hopkins Univ, Baltimore, MD 21218 USA
[3] Ohio State Univ, Columbus, OH 43210 USA
关键词
D O I
10.1007/978-3-031-43895-0_71
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep learning based medical image recognition systems often require a substantial amount of training data with expert annotations, which can be expensive and time-consuming to obtain. Recently, synthetic augmentation techniques have been proposed to mitigate the issue by generating realistic images conditioned on class labels. However, the effectiveness of these methods heavily depends on the representation capability of the trained generative model, which cannot be guaranteed without sufficient labeled training data. To further reduce the dependency on annotated data, we propose a synthetic augmentation method called HistoDiffusion, which can be pre-trained on large-scale unlabeled datasets and later applied to a small-scale labeled dataset for augmented training. In particular, we train a latent diffusion model (LDM) on diverse unlabeled datasets to learn common features and generate realistic images without conditional inputs. Then, we fine-tune the model with classifier guidance in latent space on an unseen labeled dataset so that the model can synthesize images of specific categories. Additionally, we adopt a selective mechanism to only add synthetic samples with high confidence of matching to target labels. We evaluate our proposed method by pre-training on three histopathology datasets and testing on a histopathology dataset of colorectal cancer (CRC) excluded from the pre-training datasets. With HistoDiffusion augmentation, the classification accuracy of a backbone classifier is remarkably improved by 6.4% using a small set of the original labels. Our code is available at https://github.com/karenyyy/HistoDiffAug.
引用
收藏
页码:754 / 764
页数:11
相关论文
共 50 条
  • [31] X-MOL: large-scale pre-training for molecular understanding and diverse molecular analysis
    Xue, Dongyu
    Zhang, Han
    Chen, Xiaohan
    Xiao, Dongling
    Gong, Yukang
    Chuai, Guohui
    Sun, Yu
    Tian, Hao
    Wu, Hua
    Li, Yukun
    Liu, Qi
    SCIENCE BULLETIN, 2022, 67 (09) : 899 - 902
  • [32] Score Images as a Modality: Enhancing Symbolic Music Understanding through Large-Scale Multimodal Pre-Training
    Qin, Yang
    Xie, Huiming
    Ding, Shuxue
    Li, Yujie
    Tan, Benying
    Ye, Mingchuan
    SENSORS, 2024, 24 (15)
  • [33] TeleSpeechPT: Large-Scale Chinese Multi-dialect and Multi-accent Speech Pre-training
    Chen, Hongjie
    Li, Zehan
    Xia, Guangmin
    Liu, Boqing
    Yang, Yan
    Kang, Jian
    Li, Jie
    MAN-MACHINE SPEECH COMMUNICATION, NCMMSC 2024, 2025, 2312 : 183 - 190
  • [34] EVA2.0: Investigating Open-domain Chinese Dialogue Systems with Large-scale Pre-training
    Gu, Yuxian
    Wen, Jiaxin
    Sun, Hao
    Song, Yi
    Ke, Pei
    Zheng, Chujie
    Zhang, Zheng
    Yao, Jianzhu
    Liu, Lei
    Zhu, Xiaoyan
    Huang, Minlie
    MACHINE INTELLIGENCE RESEARCH, 2023, 20 (02) : 207 - 219
  • [35] Scaling Backwards: Minimal Synthetic Pre-Training?
    Nakamura, Ryo
    Tadokoro, Ryu
    Yamada, Ryosuke
    Asano, Yuki M.
    Laina, Iro
    Rupprecht, Christian
    Inoue, Nakamasa
    Yokota, Rio
    Kataoka, Hirokatsu
    COMPUTER VISION - ECCV 2024, PT XV, 2025, 15073 : 153 - 171
  • [36] OmDet: Large-scale vision-language multi-dataset pre-training with multimodal detection network
    Zhao, Tiancheng
    Liu, Peng
    Lee, Kyusong
    IET COMPUTER VISION, 2024, 18 (05) : 626 - 639
  • [37] Efficient and Large Scale Pre-training Techniques for Japanese Natural Language Processing
    Kasagi, Akihiko
    Asaoka, Masahiro
    Tabuchi, Akihiro
    Oyama, Yosuke
    Honda, Takumi
    Sakai, Yasufumi
    Dang, Thang
    Tabaru, Tsuguchika
    2021 NINTH INTERNATIONAL SYMPOSIUM ON COMPUTING AND NETWORKING (CANDAR 2021), 2021, : 108 - 113
  • [38] HySparK: Hybrid Sparse Masking for Large Scale Medical Image Pre-training
    Tang, Fenghe
    Xu, Ronghao
    Yao, Qingsong
    Fu, Xueming
    Quan, Quan
    Zhu, Heqin
    Liu, Zaiyi
    Zhou, S. Kevin
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2024, PT XI, 2024, 15011 : 330 - 340
  • [39] CarsiDock: a deep learning paradigm for accurate protein-ligand docking and screening based on large-scale pre-training
    Cai, Heng
    Shen, Chao
    Jian, Tianye
    Zhang, Xujun
    Chen, Tong
    Han, Xiaoqi
    Yang, Zhuo
    Dang, Wei
    Hsieh, Chang-Yu
    Kang, Yu
    Pan, Peichen
    Ji, Xiangyang
    Song, Jianfei
    Hou, Tingjun
    Deng, Yafeng
    CHEMICAL SCIENCE, 2024, 15 (04) : 1449 - 1471
  • [40] Self-supervised pre-training for large-scale crop mapping using Sentinel-2 time series
    Xu, Yijia
    Ma, Yuchi
    Zhang, Zhou
    ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2024, 207 : 312 - 325