Re-GAN: Data-Efficient GANs Training via Architectural Reconfiguration

被引:14
|
作者
Saxena, Divya [1 ]
Cao, Jiannong [1 ]
Xu, Jiahao [1 ]
Kulshrestha, Tarun [1 ]
机构
[1] Hong Kong Polytech Univ, Hong Kong, Peoples R China
关键词
D O I
10.1109/CVPR52729.2023.01557
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Training Generative Adversarial Networks (GANs) on high-fidelity images usually requires a vast number of training images. Recent research on GAN tickets reveals that dense GANs models contain sparse sub-networks or lottery tickets that, when trained separately, yield better results under limited data. However, finding GANs tickets requires an expensive process of train-prune-retrain. In this paper, we propose Re-GAN, a data-efficient GANs training that dynamically reconfigures GANs architecture during training to explore different sub-network structures in training time. Our method repeatedly prunes unimportant connections to regularize GANs network and regrows them to reduce the risk of prematurely pruning important connections. Re-GAN stabilizes the GANs models with less data and offers an alternative to the existing GANs tickets and progressive growing methods. We demonstrate that Re-GAN is a generic training methodology which achieves stability on datasets of varying sizes, domains, and resolutions (CIFAR-10, Tiny-ImageNet, and multiple few-shot generation datasets) as well as different GANs architectures (SNGAN, ProGAN, StyleGAN2 and AutoGAN). Re-GAN also improves performance when combined with the recent augmentation approaches. Moreover, Re-GAN requires fewer floating-point operations (FLOPs) and less training time by removing the unimportant connections during GANs training while maintaining comparable or even generating higher-quality samples. When compared to state-of-the-art StyleGAN2, our method outperforms without requiring any additional fine-tuning step. Code can be found at this link: https://github.com/IntellicentAI-Lab/Re-GAN
引用
收藏
页码:16230 / 16240
页数:11
相关论文
共 50 条
  • [1] Data-Efficient GAN Training Beyond (Just) Augmentations: A Lottery Ticket Perspective
    Chen, Tianlong
    Cheng, Yu
    Gan, Zhe
    Liu, Jingjing
    Wang, Zhangyang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [2] NICE: NoIse-modulated Consistency rEgularization for Data-Efficient GANs
    Ni, Yao
    Koniusz, Piotr
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [3] FakeCLR: Exploring Contrastive Learning for Solving Latent Discontinuity in Data-Efficient GANs
    Li, Ziqiang
    Wang, Chaoyue
    Zheng, Heliang
    Zhang, Jing
    Li, Bin
    COMPUTER VISION - ECCV 2022, PT XV, 2022, 13675 : 598 - 615
  • [4] Data-Efficient Augmentation for Training Neural Networks
    Liu, Tian Yu
    Mirzasoleiman, Baharan
    Advances in Neural Information Processing Systems, 2022, 35
  • [5] Data-Efficient Augmentation for Training Neural Networks
    Liu, Tian Yu
    Mirzasoleiman, Baharan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [6] Data-Efficient Training Strategies for Neural TTS Systems
    Prajwal, K. R.
    Jawahar, C., V
    CODS-COMAD 2021: PROCEEDINGS OF THE 3RD ACM INDIA JOINT INTERNATIONAL CONFERENCE ON DATA SCIENCE & MANAGEMENT OF DATA (8TH ACM IKDD CODS & 26TH COMAD), 2021, : 223 - 227
  • [7] Sobolev Training for Data-efficient Approximate Nonlinear MPC
    Lueken, Lukas
    Brandner, Dean
    Lucia, Sergio
    IFAC PAPERSONLINE, 2023, 56 (02): : 5765 - 5772
  • [8] A Data-Efficient Training Method for Deep Reinforcement Learning
    Feng, Wenhui
    Han, Chongzhao
    Lian, Feng
    Liu, Xia
    ELECTRONICS, 2022, 11 (24)
  • [9] A Data-Efficient Deep Learning Strategy for Tissue Characterization via Quantitative Ultrasound: Zone Training
    Soylu, Ufuk
    Oelze, Michael L.
    IEEE TRANSACTIONS ON ULTRASONICS FERROELECTRICS AND FREQUENCY CONTROL, 2023, 70 (05) : 368 - 377
  • [10] Data-Efficient Learning via Minimizing Hyperspherical Energy
    Cao, Xiaofeng
    Liu, Weiyang
    Tsang, Ivor W.
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (11) : 13422 - 13437