Multi-Label Conditional Generation From Pre-Trained Models

被引:0
|
作者
Proszewska, Magdalena [1 ]
Wolczyk, Maciej [1 ]
Zieba, Maciej [2 ,3 ]
Wielopolski, Patryk [4 ]
Maziarka, Lukasz [1 ]
Smieja, Marek [1 ]
机构
[1] Jagiellonian Univ, Fac Math & Comp Sci, PL-31007 Krakow, Poland
[2] Tooploox, PL-53601 Wroclaw, Poland
[3] Wroclaw Univ Sci & Technol, PL-53601 Wroclaw, Poland
[4] Wroclaw Univ Sci & Technol, PL-50370 Wroclaw, Poland
关键词
Training; Computational modeling; Adaptation models; Vectors; Data models; Aerospace electronics; Three-dimensional displays; Conditional generation; deep generative models; GANs; invertible normalizing flows; pre-trained models; VAEs;
D O I
10.1109/TPAMI.2024.3382008
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Although modern generative models achieve excellent quality in a variety of tasks, they often lack the essential ability to generate examples with requested properties, such as the age of the person in the photo or the weight of the generated molecule. To overcome these limitations we propose PluGeN (Plugin Generative Network), a simple yet effective generative technique that can be used as a plugin for pre-trained generative models. The idea behind our approach is to transform the entangled latent representation using a flow-based module into a multi-dimensional space where the values of each attribute are modeled as an independent one-dimensional distribution. In consequence, PluGeN can generate new samples with desired attributes as well as manipulate labeled attributes of existing examples. Due to the disentangling of the latent representation, we are even able to generate samples with rare or unseen combinations of attributes in the dataset, such as a young person with gray hair, men with make-up, or women with beards. In contrast to competitive approaches, PluGeN can be trained on partially labeled data. We combined PluGeN with GAN and VAE models and applied it to conditional generation and manipulation of images, chemical molecule modeling and 3D point clouds generation.
引用
收藏
页码:6185 / 6198
页数:14
相关论文
共 50 条
  • [41] DistillingWord Meaning in Context from Pre-trained Language Models
    Arase, Yuki
    Kajiwara, Tomoyuki
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 534 - 546
  • [42] An Investigation of Suitability of Pre-Trained Language Models for Dialogue Generation - Avoiding Discrepancies
    Zeng, Yan
    Nie, Jian-Yun
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 4481 - 4494
  • [43] Addressing Extraction and Generation Separately: Keyphrase Prediction With Pre-Trained Language Models
    Liu, Rui
    Lin, Zheng
    Wang, Weiping
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2021, 29 : 3180 - 3191
  • [44] On the generation of multi-label prototypes
    Bello, Marilyn
    Napoles, Gonzalo
    Vanhoof, Koen
    Bello, Rafael
    INTELLIGENT DATA ANALYSIS, 2020, 24 (S1) : S167 - S183
  • [45] Automatic Title Generation for Learning Resources and Pathways with Pre-trained Transformer Models
    Mishra, Prakhar
    Diwan, Chaitali
    Srinivasa, Srinath
    Srinivasaraghavan, G.
    INTERNATIONAL JOURNAL OF SEMANTIC COMPUTING, 2021, 15 (04) : 487 - 510
  • [46] Multi-Label Clinical Time-Series Generation via Conditional GAN
    Lu, Chang
    Reddy, Chandan K.
    Wang, Ping
    Nie, Dong
    Ning, Yue
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (04) : 1728 - 1740
  • [47] Annotating Columns with Pre-trained Language Models
    Suhara, Yoshihiko
    Li, Jinfeng
    Li, Yuliang
    Zhang, Dan
    Demiralp, Cagatay
    Chen, Chen
    Tan, Wang-Chiew
    PROCEEDINGS OF THE 2022 INTERNATIONAL CONFERENCE ON MANAGEMENT OF DATA (SIGMOD '22), 2022, : 1493 - 1503
  • [48] MtArtGPT: A Multi-Task Art Generation System With Pre-Trained Transformer
    Jin, Cong
    Zhu, Ruolin
    Zhu, Zixing
    Yang, Lu
    Yang, Min
    Luo, Jiebo
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (08) : 6901 - 6912
  • [49] Lottery Jackpots Exist in Pre-Trained Models
    Zhang, Yuxin
    Lin, Mingbao
    Zhong, Yunshan
    Chao, Fei
    Ji, Rongrong
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (12) : 14990 - 15004
  • [50] Interpreting Art by Leveraging Pre-Trained Models
    Penzel, Niklas
    Denzler, Joachim
    2023 18TH INTERNATIONAL CONFERENCE ON MACHINE VISION AND APPLICATIONS, MVA, 2023,