GANtlitz: Ultra High Resolution Generative Model for Multi-Modal Face Textures

被引:0
|
作者
Gruber, A. [1 ,2 ]
Collins, E. [2 ]
Meka, A. [2 ]
Mueller, F. [2 ]
Sarkar, K. [2 ]
Orts-Escolano, S. [2 ]
Prasso, L. [2 ]
Busch, J. [2 ]
Gross, M. [1 ]
Beeler, T. [2 ]
机构
[1] ETH, Zurich, Switzerland
[2] Google, Menlo Pk, CA USA
关键词
<bold>CCS Concepts</bold>; center dot <bold>Computing methodologies</bold> -> <bold>Machine learning</bold>; <bold>Texturing</bold>;
D O I
10.1111/cgf.15039
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
High-resolution texture maps are essential to render photoreal digital humans for visual effects or to generate data for machine learning. The acquisition of high resolution assets at scale is cumbersome, it involves enrolling a large number of human subjects, using expensive multi-view camera setups, and significant manual artistic effort to align the textures. To alleviate these problems, we introduce GANtlitz (A play on the german noun Antlitz, meaning face), a generative model that can synthesize multi-modal ultra-high-resolution face appearance maps for novel identities. Our method solves three distinct challenges: 1) unavailability of a very large data corpus generally required for training generative models, 2) memory and computational limitations of training a GAN at ultra-high resolutions, and 3) consistency of appearance features such as skin color, pores and wrinkles in high-resolution textures across different modalities. We introduce dual-style blocks, an extension to the style blocks of the StyleGAN2 architecture, which improve multi-modal synthesis. Our patch-based architecture is trained only on image patches obtained from a small set of face textures (<100) and yet allows us to generate seamless appearance maps of novel identities at 6k x 4k resolution. Extensive qualitative and quantitative evaluations and baseline comparisons show the efficacy of our proposed system. (see )
引用
收藏
页数:14
相关论文
共 50 条
  • [31] Face Recognition using Multi-modal Binary Patterns
    Thanh Phuong Nguyen
    Ngoc-Son Vu
    Caplier, Alice
    2012 21ST INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR 2012), 2012, : 2343 - 2346
  • [32] An Overview of Multi-Modal Biometrics Based on Face and Ear
    Zhang, Haijun
    Huang, Zengxi
    Li, Yibo
    2009 IEEE INTERNATIONAL CONFERENCE ON AUTOMATION AND LOGISTICS ( ICAL 2009), VOLS 1-3, 2009, : 1705 - 1709
  • [33] Multi-modal face tracking using Bayesian network
    Liu, F
    Lin, XY
    Lie, SZ
    Shi, YC
    IEEE INTERNATIONAL WORKSHOP ON ANALYSIS AND MODELING OF FACE AND GESTURES, 2003, : 135 - 142
  • [34] Collaborative Diffusion for Multi-Modal Face Generation and Editing
    Huang, Ziqi
    Chan, Kelvin C. K.
    Jiang, Yuming
    Liu, Ziwei
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 6080 - 6090
  • [35] Research and Implementation of of Multi-modal Face Recognition Algorithm
    Ye Jihua
    Xia Guomiao
    Hu Dan
    2013 25TH CHINESE CONTROL AND DECISION CONFERENCE (CCDC), 2013, : 2086 - 2090
  • [36] Towards a multi-modal perceptual model
    Hollier, MP
    Voelcker, R
    BT TECHNOLOGY JOURNAL, 1997, 15 (04): : 162 - 171
  • [37] Multi-modal Background Model Initialization
    Bloisi, Domenico D.
    Grillo, Alfonso
    Pennisi, Andrea
    Iocchi, Luca
    Passaretti, Claudio
    NEW TRENDS IN IMAGE ANALYSIS AND PROCESSING - ICIAP 2015 WORKSHOPS, 2015, 9281 : 485 - 492
  • [38] HMGAN: A Hierarchical Multi-Modal Generative Adversarial Network Model for Wearable Human Activity Recognition
    Chen, Ling
    Hu, Rong
    Wu, Menghan
    Zhou, Xin
    PROCEEDINGS OF THE ACM ON INTERACTIVE MOBILE WEARABLE AND UBIQUITOUS TECHNOLOGIES-IMWUT, 2023, 7 (03):
  • [39] A hierarchical multi-modal cross-attention model for face anti-spoofing
    Xue, Hao
    Ma, Jing
    Guo, Xiaoyu
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2023, 97