Interactive Generative Adversarial Networks With High-Frequency Compensation for Facial Attribute Editing

被引:0
|
作者
Huang, Wenmin [1 ,2 ]
Luo, Weiqi [1 ,2 ]
Cao, Xiaochun [3 ]
Huang, Jiwu [4 ]
机构
[1] Sun Yat Sen Univ, Guangdong Prov Key Lab Informat Secur Technol, Guangzhou 510000, Guangdong, Peoples R China
[2] Sun Yat Sen Univ, Sch Comp Sci & Engn, Guangzhou 510000, Guangdong, Peoples R China
[3] Sun Yat Sen Univ, Sch Cyber Sci & Technol, Shenzhen Campus, Shenzhen 518107, Peoples R China
[4] Shenzhen MSU BIT Univ, Fac Engn, Guangdong Lab Machine Percept & Intelligent Comp, Shenzhen 518116, Peoples R China
基金
中国国家自然科学基金;
关键词
Facial attribute editing; cross-task interaction; generative adversarial network;
D O I
10.1109/TCSVT.2024.3391348
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Recently, facial attribute editing has drawn increasing attention and has achieved significant progress due to Generative Adversarial Network (GAN). Since paired images before and after editing are not available, existing methods typically perform the editing and reconstruction tasks simultaneously, and transfer facial details learned from the reconstruction to the editing via sharing the latent representation space and weights. In this way, they can not preserve those non-targeted regions well during editing. In addition, they usually introduce skip connections between the encoder and decoder to improve image quality at the cost of attribute editing ability. In this paper, we propose a novel method called InterGAN with high-frequency compensation to alleviate above problems. Specifically, we first propose the cross-task interaction (CTI) to fully explore the relationships between editing and reconstruction tasks. The CTI includes two translations: style translation adjusts the mean and variance of feature maps according to style features, and conditional translation utilizes attribute vector as condition to guide feature map transformation. They provide effective information interaction to preserve the irrelevant regions unchanged. Without using skip connections between the encoder and decoder, furthermore, we propose the high-frequency compensation module (HFCM) to improve image quality. The HFCM tries to collect potentially loss information from input images and each down-sampling layers of the encoder, and then re-inject them into subsequent layers to alleviate the information loss. Ablation analysis show the effectiveness of proposed CTI and HFCM. Extensive qualitative and quantitative experiments on CelebA-HQ demonstrate that the proposed method outperforms state-of-the-art methods both in attribute editing accuracy and image quality.
引用
收藏
页码:8215 / 8229
页数:15
相关论文
共 50 条
  • [1] Face attribute editing based on generative adversarial networks
    Xiaoxia Song
    Mingwen Shao
    Wangmeng Zuo
    Cunhe Li
    Signal, Image and Video Processing, 2020, 14 : 1217 - 1225
  • [2] Face attribute editing based on generative adversarial networks
    Song, Xiaoxia
    Shao, Mingwen
    Zuo, Wangmeng
    Li, Cunhe
    SIGNAL IMAGE AND VIDEO PROCESSING, 2020, 14 (06) : 1217 - 1225
  • [3] Progressive editing with stacked Generative Adversarial Network for multiple facial attribute editing
    Chan, Patrick P. K.
    Wang, Xiaotian
    Lin, Zhe
    Yeung, Daniel S.
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2022, 217
  • [4] InjectionGAN: Unified Generative Adversarial Networks for Arbitrary Image Attribute Editing
    Ding, Chen
    Kang, Wei
    Zhu, Jiaqi
    Du, Shuangyan
    IEEE ACCESS, 2020, 8 : 117726 - 117735
  • [5] Semi-supervised image attribute editing using generative adversarial networks
    Dogan, Yahya
    Keles, Hacer Yalim
    NEUROCOMPUTING, 2020, 401 (401) : 338 - 352
  • [6] Facial attribute-controlled sketch-to-image translation with generative adversarial networks
    Hu, Mingming
    Guo, Jingtao
    EURASIP JOURNAL ON IMAGE AND VIDEO PROCESSING, 2020, 2020 (01)
  • [7] Facial attribute-controlled sketch-to-image translation with generative adversarial networks
    Mingming Hu
    Jingtao Guo
    EURASIP Journal on Image and Video Processing, 2020
  • [8] Sparsely Grouped Multi-Task Generative Adversarial Networks for Facial Attribute Manipulation
    Zhang, Jichao
    Shu, Yezhi
    Xu, Songhua
    Cao, Gongze
    Zhong, Fan
    Liu, Meng
    Qin, Xueying
    PROCEEDINGS OF THE 2018 ACM MULTIMEDIA CONFERENCE (MM'18), 2018, : 392 - 401
  • [9] Generative Adversarial Network with Spatial Attention for Face Attribute Editing
    Zhang, Gang
    Kan, Meina
    Shan, Shiguang
    Chen, Xilin
    COMPUTER VISION - ECCV 2018, PT VI, 2018, 11210 : 422 - 437
  • [10] Generative Attribute Controller with Conditional Filtered Generative Adversarial Networks
    Kaneko, Takuhiro
    Hiramatsu, Kaoru
    Kashino, Kunio
    30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 7006 - 7015