PluGeN: Multi-Label Conditional Generation from Pre-trained Models

被引:0
|
作者
Wolczyk, Maciej [1 ]
Proszewska, Magdalena [1 ]
Maziarka, Lukasz [1 ]
Zieba, Maciej [2 ,4 ]
Wielopolski, Patryk [2 ]
Kurczab, Rafal [3 ]
Smieja, Marek [1 ]
机构
[1] Jagiellonian Univ, Krakow, Poland
[2] Wroclaw Univ Sci & Technol, Wroclaw, Poland
[3] Inst Pharmacol PAS, Warsaw, Poland
[4] Tooploox, Wroclaw, Poland
关键词
REPRESENTATION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Modern generative models achieve excellent quality in a variety of tasks including image or text generation and chemical molecule modeling. However, existing methods often lack the essential ability to generate examples with requested properties, such as the age of the person in the photo or the weight of the generated molecule. Incorporating such additional conditioning factors would require rebuilding the entire architecture and optimizing the parameters from scratch. Moreover, it is difficult to disentangle selected attributes so that to perform edits of only one attribute while leaving the others unchanged. To overcome these limitations we propose PluGeN (Plugin Generative Network), a simple yet effective generative technique that can be used as a plugin to pre-trained generative models. The idea behind our approach is to transform the entangled latent representation using a flow-based module into a multi-dimensional space where the values of each attribute are modeled as an independent one-dimensional distribution. In consequence, PluGeN can generate new samples with desired attributes as well as manipulate labeled attributes of existing examples. Due to the disentangling of the latent representation, we are even able to generate samples with rare or unseen combinations of attributes in the dataset, such as a young person with gray hair, men with make-up, or women with beards. We combined PluGeN with GAN and VAE models and applied it to conditional generation and manipulation of images and chemical molecule modeling. Experiments demonstrate that PluGeN preserves the quality of backbone models while adding the ability to control the values of labeled attributes. Implementation is available at https://github.com/gmum/plugen.
引用
收藏
页码:8647 / 8656
页数:10
相关论文
共 50 条
  • [1] Multi-Label Conditional Generation From Pre-Trained Models
    Proszewska, Magdalena
    Wolczyk, Maciej
    Zieba, Maciej
    Wielopolski, Patryk
    Maziarka, Lukasz
    Smieja, Marek
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (09) : 6185 - 6198
  • [2] EmoBART: Multi-label Emotion Classification Method Based on Pre-trained Label Sequence Generation Model
    Chen, Sufen
    Chen, Lei
    Zeng, Xuegiang
    NEURAL COMPUTING FOR ADVANCED APPLICATIONS, NCAA 2024, PT III, 2025, 2183 : 104 - 115
  • [3] Ensembling Multilingual Pre-Trained Models for Predicting Multi-Label Regression Emotion Share from Speech
    Atmaja, Bagus Tris
    Sasou, Akira
    2023 ASIA PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE, APSIPA ASC, 2023, : 1026 - 1029
  • [4] Single-label and multi-label conceptor classifiers in pre-trained neural networks
    Qian, Guangwu
    Zhang, Lei
    Wang, Yan
    NEURAL COMPUTING & APPLICATIONS, 2019, 31 (10): : 6179 - 6188
  • [5] Single-label and multi-label conceptor classifiers in pre-trained neural networks
    Guangwu Qian
    Lei Zhang
    Yan Wang
    Neural Computing and Applications, 2019, 31 : 6179 - 6188
  • [6] Leveraging Pre-Trained Extreme Multi-Label Classifiers for Zero-Shot Learning
    Ostapuk, Natalia
    Dolamic, Ljiljana
    Mermoud, Alain
    Cudre-Mauroux, Philippe
    2024 11TH IEEE SWISS CONFERENCE ON DATA SCIENCE, SDS 2024, 2024, : 233 - 236
  • [7] Transfer learning with pre-trained conditional generative models
    Yamaguchi, Shin'ya
    Kanai, Sekitoshi
    Kumagai, Atsutoshi
    Chijiwa, Daiki
    Kashima, Hisashi
    MACHINE LEARNING, 2025, 114 (04)
  • [8] Pseudo-Prompt Generating in Pre-trained Vision-Language Models for Multi-label Medical Image Classification
    Ye, Yaoqin
    Zhang, Junjie
    Shi, Hongwei
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2024, PT XIV, 2025, 15044 : 279 - 298
  • [9] Research on cross-lingual multi-label patent classification based on pre-trained model
    Lu, Yonghe
    Chen, Lehua
    Tong, Xinyu
    Peng, Yongxin
    Zhu, Hou
    SCIENTOMETRICS, 2024, 129 (06) : 3067 - 3087
  • [10] Conditional pre-trained attention based Chinese question generation
    Zhang, Liang
    Fang, Ligang
    Fan, Zheng
    Li, Wei
    An, Jing
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2021, 33 (20):