PluGeN: Multi-Label Conditional Generation from Pre-trained Models

被引:0
|
作者
Wolczyk, Maciej [1 ]
Proszewska, Magdalena [1 ]
Maziarka, Lukasz [1 ]
Zieba, Maciej [2 ,4 ]
Wielopolski, Patryk [2 ]
Kurczab, Rafal [3 ]
Smieja, Marek [1 ]
机构
[1] Jagiellonian Univ, Krakow, Poland
[2] Wroclaw Univ Sci & Technol, Wroclaw, Poland
[3] Inst Pharmacol PAS, Warsaw, Poland
[4] Tooploox, Wroclaw, Poland
关键词
REPRESENTATION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Modern generative models achieve excellent quality in a variety of tasks including image or text generation and chemical molecule modeling. However, existing methods often lack the essential ability to generate examples with requested properties, such as the age of the person in the photo or the weight of the generated molecule. Incorporating such additional conditioning factors would require rebuilding the entire architecture and optimizing the parameters from scratch. Moreover, it is difficult to disentangle selected attributes so that to perform edits of only one attribute while leaving the others unchanged. To overcome these limitations we propose PluGeN (Plugin Generative Network), a simple yet effective generative technique that can be used as a plugin to pre-trained generative models. The idea behind our approach is to transform the entangled latent representation using a flow-based module into a multi-dimensional space where the values of each attribute are modeled as an independent one-dimensional distribution. In consequence, PluGeN can generate new samples with desired attributes as well as manipulate labeled attributes of existing examples. Due to the disentangling of the latent representation, we are even able to generate samples with rare or unseen combinations of attributes in the dataset, such as a young person with gray hair, men with make-up, or women with beards. We combined PluGeN with GAN and VAE models and applied it to conditional generation and manipulation of images and chemical molecule modeling. Experiments demonstrate that PluGeN preserves the quality of backbone models while adding the ability to control the values of labeled attributes. Implementation is available at https://github.com/gmum/plugen.
引用
收藏
页码:8647 / 8656
页数:10
相关论文
共 50 条
  • [11] DeBERTa-BiLSTM: A multi-label classification model of Arabic medical questions using pre-trained models and deep learning
    Al-Smadi, Bushra Salem
    COMPUTERS IN BIOLOGY AND MEDICINE, 2024, 170
  • [12] Comparison of pre-trained language models in terms of carbon emissions, time and accuracy in multi-label text classification using AutoML
    Savci, Pinar
    Das, Bihter
    HELIYON, 2023, 9 (05)
  • [13] Multi-label Fine-Grained Entity Typing for Baidu Wikipedia Based on Pre-trained Model
    Pu, Keyu
    Liu, Hongyi
    Yang, Yixiao
    Lv, Wenyi
    Li, Jinlong
    CCKS 2021 - EVALUATION TRACK, 2022, 1553 : 114 - 123
  • [14] TED TALK TEASER GENERATION WITH PRE-TRAINED MODELS
    Vico, Gianluca
    Niehues, Jan
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 8067 - 8071
  • [15] Pre-Trained Language Models for Text Generation: A Survey
    Li, Junyi
    Tang, Tianyi
    Zhao, Wayne Xin
    Nie, Jian-Yun
    Wen, Ji-Rong
    ACM COMPUTING SURVEYS, 2024, 56 (09)
  • [16] Leveraging pre-trained language models for code generation
    Soliman, Ahmed
    Shaheen, Samir
    Hadhoud, Mayada
    COMPLEX & INTELLIGENT SYSTEMS, 2024, 10 (03) : 3955 - 3980
  • [17] Neural Conditional Energy Models for Multi-Label Classification
    Jing, How
    Lin, Shou-De
    2014 IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2014, : 240 - 249
  • [18] Controllable Generation from Pre-trained Language Models via Inverse Prompting
    Zou, Xu
    Yin, Da
    Zhong, Qingyang
    Yang, Hongxia
    Yang, Zhilin
    Tang, Jie
    KDD '21: PROCEEDINGS OF THE 27TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2021, : 2450 - 2460
  • [19] Attribute Alignment: Controlling Text Generation from Pre-trained Language Models
    Yu, Dian
    Yu, Zhou
    Sagae, Kenji
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 2251 - 2268
  • [20] Adaptive Textual Label Noise Learning based on Pre-trained Models
    Cheng, Shaohuan
    Chen, Wenyu
    Fu, Mingsheng
    Xie, Xuanting
    Qu, Hong
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 3174 - 3188