GANtlitz: Ultra High Resolution Generative Model for Multi-Modal Face Textures

被引:0
|
作者
Gruber, A. [1 ,2 ]
Collins, E. [2 ]
Meka, A. [2 ]
Mueller, F. [2 ]
Sarkar, K. [2 ]
Orts-Escolano, S. [2 ]
Prasso, L. [2 ]
Busch, J. [2 ]
Gross, M. [1 ]
Beeler, T. [2 ]
机构
[1] ETH, Zurich, Switzerland
[2] Google, Menlo Pk, CA USA
关键词
<bold>CCS Concepts</bold>; center dot <bold>Computing methodologies</bold> -> <bold>Machine learning</bold>; <bold>Texturing</bold>;
D O I
10.1111/cgf.15039
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
High-resolution texture maps are essential to render photoreal digital humans for visual effects or to generate data for machine learning. The acquisition of high resolution assets at scale is cumbersome, it involves enrolling a large number of human subjects, using expensive multi-view camera setups, and significant manual artistic effort to align the textures. To alleviate these problems, we introduce GANtlitz (A play on the german noun Antlitz, meaning face), a generative model that can synthesize multi-modal ultra-high-resolution face appearance maps for novel identities. Our method solves three distinct challenges: 1) unavailability of a very large data corpus generally required for training generative models, 2) memory and computational limitations of training a GAN at ultra-high resolutions, and 3) consistency of appearance features such as skin color, pores and wrinkles in high-resolution textures across different modalities. We introduce dual-style blocks, an extension to the style blocks of the StyleGAN2 architecture, which improve multi-modal synthesis. Our patch-based architecture is trained only on image patches obtained from a small set of face textures (<100) and yet allows us to generate seamless appearance maps of novel identities at 6k x 4k resolution. Extensive qualitative and quantitative evaluations and baseline comparisons show the efficacy of our proposed system. (see )
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Multi-Modal Face Stylization with a Generative Prior
    Li, Mengtian
    Dong, Yi
    Lin, Minxuan
    Huang, Haibin
    Wan, Pengfei
    Ma, Chongyang
    COMPUTER GRAPHICS FORUM, 2023, 42 (07)
  • [2] Multi-Modal Face Recognition
    Shen, Haihong
    Ma, Liqun
    Zhang, Qishan
    2ND IEEE INTERNATIONAL CONFERENCE ON ADVANCED COMPUTER CONTROL (ICACC 2010), VOL. 5, 2010, : 612 - 616
  • [3] Multi-Modal Face Recognition
    Shen, Haihong
    Ma, Liqun
    Zhang, Qishan
    2010 8TH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION (WCICA), 2010, : 720 - 723
  • [4] A Generative Model for Brain Tumor Segmentation in Multi-Modal Images
    Menze, Bjoern H.
    Van Leemput, Koen
    Lashkari, Danial
    Weber, Marc-Andre
    Ayache, Nicholas
    Golland, Polina
    MEDICAL IMAGE COMPUTING AND COMPUTER-ASSISTED INTERVENTION - MICCAI 2010, PT II,, 2010, 6362 : 151 - +
  • [5] Multi-modal tensor face for simultaneous super-resolution and recognition
    Jia, K
    Gong, SG
    TENTH IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION, VOLS 1 AND 2, PROCEEDINGS, 2005, : 1683 - 1690
  • [6] Multi-Modal Generative AI with Foundation Models
    Liu, Ziwei
    PROCEEDINGS OF THE 2ND WORKSHOP ON LARGE GENERATIVE MODELS MEET MULTIMODAL APPLICATIONS, LGM(CUBE)A 2024, 2024, : 4 - 4
  • [7] Discriminative multi-modal deep generative models
    Du, Fang
    Zhang, Jiangshe
    Hu, Junying
    Fei, Rongrong
    KNOWLEDGE-BASED SYSTEMS, 2019, 173 : 74 - 82
  • [8] Multi-Modal Generative AI with Foundation Models
    Liu, Ziwei
    PROCEEDINGS OF THE 1ST WORKSHOP ON LARGE GENERATIVE MODELS MEET MULTIMODAL APPLICATIONS, LGM3A 2023, 2023, : 5 - 5
  • [9] Multi-modal ultra-high resolution structural 7-Tesla MRI data repository
    Forstmann, Birte U.
    Keuken, Max C.
    Schafer, Andreas
    Bazin, Pierre-Louis
    Alkemade, Anneke
    Turner, Robert
    SCIENTIFIC DATA, 2014, 1
  • [10] Multi-modal ultra-high resolution structural 7-Tesla MRI data repository
    Birte U Forstmann
    Max C Keuken
    Andreas Schafer
    Pierre-Louis Bazin
    Anneke Alkemade
    Robert Turner
    Scientific Data, 1