StyleAutoEncoder for Manipulating Image Attributes Using Pre-trained StyleGAN

被引:0
|
作者
Bedychaj, Andrzej [1 ]
Tabor, Jacek [1 ]
Smieja, Marek [1 ]
机构
[1] Jagiellonian Univ, Krakow, Poland
关键词
D O I
10.1007/978-981-97-2253-2_10
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep conditional generative models are excellent tools for creating high-quality images and editing their attributes. However, training modern generative models from scratch is very expensive and requires large computational resources. In this paper, we introduce StyleAutoEncoder (StyleAE), a lightweight AutoEncoder module, which works as a plugin for pre-trained generative models and allows for manipulating the requested attributes of images. The proposed method offers a cost-effective solution for training deep generative models with limited computational resources, making it a promising technique for a wide range of applications. We evaluate StyleAE by combining it with StyleGAN, which is currently one of the top generative models. Our experiments demonstrate that StyleAE is at least as effective in manipulating image attributes as the state-of-the-art algorithms based on invertible normalizing flows. However, it is simpler, faster, and gives more freedom in designing neural architecture.
引用
收藏
页码:118 / 130
页数:13
相关论文
共 50 条
  • [31] Calibration of Pre-trained Transformers
    Desai, Shrey
    Durrett, Greg
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 295 - 302
  • [32] Attacking Pre-trained Recommendation
    Wu, Yiqing
    Xie, Ruobing
    Zhang, Zhao
    Zhu, Yongchun
    Zhuang, Fuzhen
    Zhou, Jie
    Xu, Yongjun
    He, Qing
    PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, : 1811 - 1815
  • [33] Pre-trained Adversarial Perturbations
    Ban, Yuanhao
    Dong, Yinpeng
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [34] Improved Arabic image captioning model using feature concatenation with pre-trained word embedding
    Elbedwehy, Samar
    Medhat, T.
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (26): : 19051 - 19067
  • [35] Optimization of a pre-trained AlexNet model for detecting and localizing image forgeries
    Samir S.
    Emary E.
    El-Sayed K.
    Onsi H.
    Samir, Soad (s.samir@fci-cu.edu.eg), 1600, MDPI AG (11):
  • [36] Adapting Pre-trained Generative Model to Medical Image for Data Augmentation
    Yuan, Zhouhang
    Fang, Zhengqing
    Huang, Zhengxing
    Wu, Fei
    Yao, Yu-Feng
    Li, Yingming
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2024, PT V, 2024, 15005 : 79 - 89
  • [37] Video Colorization with Pre-trained Text-to-Image Diffusion Models
    Liu, Hanyuan
    Xie, Minshan
    Xing, Jinbo
    Li, Chengze
    Wong, Tien-Tsin
    arXiv, 2023,
  • [38] Using BERT pre-trained image transformers to identify potential parametric Wafer Map Defects
    Jen, En
    Ting, Yiju
    Chen, Boris
    Jan, C. H.
    Huang, Lester
    Lin, ChingYu
    Wu, Milton
    Feng, Anna
    Wen, Charless
    Chen, H. W.
    Yeh, Jason
    Lai, Citi
    2024 35TH ANNUAL SEMI ADVANCED SEMICONDUCTOR MANUFACTURING CONFERENCE, ASMC, 2024,
  • [39] Hybrid image super-resolution using perceptual similarity from pre-trained network
    Chen, Yanxiang
    Tan, Huadong
    Zhang, Luming
    Zhou, Jie
    Lu, Qiang
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2019, 60 : 229 - 235
  • [40] SVBRDF Recovery from a Single Image with Highlights Using a Pre-trained Generative Adversarial Network
    Wen, Tao
    Wang, Beibei
    Zhang, Lei
    Guo, Jie
    Holzschuch, Nicolas
    COMPUTER GRAPHICS FORUM, 2022, 41 (06) : 110 - 123