Boosted GAN with Semantically Interpretable Information for Image Inpainting

被引:7
|
作者
Li, Ang [1 ]
Qi, Jianzhong [1 ]
Zhang, Rui [1 ]
Kotagiri, Ramamohanarao [1 ]
机构
[1] Univ Melbourne, Melbourne, Vic, Australia
关键词
image inpainting; GAN; semantic information; image attribute; image segmentation;
D O I
10.1109/ijcnn.2019.8851926
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Image inpainting aims at restoring missing regions of corrupted images, which has many applications such as image restoration and object removal. However, current GAN-based inpainting models fail to explicitly consider the semantic consistency between restored images and original images. For example, given a male image with image region of one eye missing, current models may restore it with a female eye. This is due to the ambiguity of GAN-based inpainting models: these models can generate many possible restorations given a missing region. To address this limitation, our key insight is that semantically interpretable information (such as attribute and segmentation information) of input images (with missing regions) can provide essential guidance for the inpainting process. Based on this insight, we propose a boosted GAN with semantically interpretable information for image inpainting that consists of an inpainting network and a discriminative network. The inpainting network utilizes two auxiliary pretrained networks to discover the attribute and segmentation information of input images and incorporates them into the inpainting process to provide explicit semantic-level guidance. The discriminative network adopts a multi-level design that can enforce regularizations not only on overall realness but also on attribute and segmentation consistency with the original images. Experimental results show that our proposed model can preserve consistency on both attribute and segmentation level, and significantly outperforms the state-of-the-art models.
引用
收藏
页数:8
相关论文
共 50 条
  • [31] A sparsity image inpainting algorithm combining color with gradient information
    Li, Zhidan
    He, Hongjie
    Yin, Zhongke
    Chen, Fan
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2014, 51 (09): : 2081 - 2093
  • [32] Pyramid-VAE-GAN: Transferring hierarchical latent variables for image inpainting
    Huiyuan Tian
    Li Zhang
    Shijian Li
    Min Yao
    Gang Pan
    Computational Visual Media, 2023, 9 : 827 - 841
  • [33] Preliminary Investigation on Single Remote Sensing Image Inpainting through a Modified GAN
    Lou, Shenlong
    Fan, Qiancong
    Chen, Feng
    Wang, Cheng
    Li, Jonathan
    2018 10TH IAPR WORKSHOP ON PATTERN RECOGNITION IN REMOTE SENSING (PRRS), 2018,
  • [34] Thangka image inpainting using adjacent information of broken area
    Liu, Huaming
    Wang, Weilan
    Xie, Hui
    IMECS 2008: INTERNATIONAL MULTICONFERENCE OF ENGINEERS AND COMPUTER SCIENTISTS, VOLS I AND II, 2008, : 646 - 649
  • [35] Image Inpainting with LS-SVM Based on Gradient Information
    Ji, Ruirui
    Wu, Yi
    Qi, Kaijie
    2018 CHINESE AUTOMATION CONGRESS (CAC), 2018, : 4050 - 4054
  • [36] AN EDGE INFORMATION AND MASK SHRINKING BASED IMAGE INPAINTING APPROACH
    Xu, Huali
    Su, Xiangdong
    Wang, Meng
    Hao, Xiang
    Gao, Guanglai
    2020 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2020,
  • [37] Pyramid-VAE-GAN: Transferring hierarchical latent variables for image inpainting
    Tian, Huiyuan
    Zhang, Li
    Li, Shijian
    Yao, Min
    Pan, Gang
    COMPUTATIONAL VISUAL MEDIA, 2023, 9 (04) : 827 - 841
  • [38] Multi-scale patch-GAN with edge detection for image inpainting
    Chen, Gang
    Zhang, Guipeng
    Yang, Zhenguo
    Liu, Wenyin
    APPLIED INTELLIGENCE, 2023, 53 (04) : 3917 - 3932
  • [39] InViT: GAN Inversion-Based Vision Transformer for Blind Image Inpainting
    Du, Yongqiang
    Liu, Haoran
    He, Shengjie
    Chen, Songnan
    IEEE ACCESS, 2024, 12 : 129956 - 129965
  • [40] Image inpainting
    Takahashi, Tomohiro
    Kyokai Joho Imeji Zasshi/Journal of the Institute of Image Information and Television Engineers, 2017, 71 (07): : 503 - 504