Depth-Aware Generative Adversarial Network for Talking Head Video Generation

被引:64
|
作者
Hong, Fa-Ting [1 ]
Zhang, Longhao [2 ]
Shen, Li [2 ]
Xu, Dan [1 ]
机构
[1] HKUST, Dept Comp Sci & Engn, Hong Kong, Peoples R China
[2] Alibaba Cloud, Hangzhou, Peoples R China
关键词
D O I
10.1109/CVPR52688.2022.00339
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Talking head video generation aims to produce a synthetic human face video that contains the identity and pose information respectively from a given source image and a driving video. Existing works for this task heavily rely on 2D representations (e.g. appearance and motion) learned from the input images. However, dense 3D facial geometry (e.g. pixel-wise depth) is extremely important for this task as it is particularly beneficial for us to essentially generate accurate 3D face structures and distinguish noisy information from the possibly cluttered background. Nevertheless, dense 3D geometry annotations are prohibitively costly for videos and are typically not available for this video generation task. In this paper, we introduce a self-supervised face-depth learning method to automatically recover dense 3D facial geometry (i.e. depth) from the face videos without the requirement of any expensive 31) annotation data. Based on the learned dense depth maps, we further propose to leverage them to estimate sparse facial keypoints that capture the critical movement of the human head. In a more dense way, the depth is also utilized to learn 3D-aware cross-modal (i.e. appearance and depth) attention to guide the generation of motion fields for warping source image representations. All these contributions compose a novel depth-aware generative adversarial network (DaGAN) for talking head generation. Extensive experiments conducted demonstrate that our proposed method can generate highly realistic faces, and achieve significant results on the unseen human faces.(1)
引用
收藏
页码:3387 / 3396
页数:10
相关论文
共 50 条
  • [1] DaGAN plus plus : Depth-Aware Generative Adversarial Network for Talking Head Video Generation
    Hong, Fa-Ting
    Shen, Li
    Xu, Dan
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (05) : 2997 - 3012
  • [2] Depth-Aware Stereo Video Retargeting
    Li, Bing
    Lin, Chia-Wen
    Shi, Boxin
    Huang, Tiejun
    Gao, Wen
    Kuo, C. -C. Jay
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 6517 - 6525
  • [3] Depth-Aware Unpaired Video Dehazing
    Yang, Yang
    Guo, Chun-Le
    Guo, Xiaojie
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 (2388-2403) : 2388 - 2403
  • [4] Depth-Aware Video Frame Interpolation
    Bao, Wenbo
    Lai, Wei-Sheng
    Ma, Chao
    Zhang, Xiaoyun
    Gao, Zhiyong
    Yang, Ming-Hsuan
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 3698 - 3707
  • [5] Depth-Aware Endoscopic Video Inpainting
    Zhang, Francis Xiatian
    Chen, Shuang
    Xie, Xianghua
    Shum, Hubert P. H.
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2024, PT VI, 2024, 15006 : 143 - 153
  • [6] Depth-Aware Image Colorization Network
    Chu, Wei-Ta
    Hsu, Yu-Ting
    PROCEEDINGS OF THE 2018 WORKSHOP ON UNDERSTANDING SUBJECTIVE ATTRIBUTES OF DATA, WITH THE FOCUS ON EVOKED EMOTIONS (EE-USAD'18), 2018, : 17 - 23
  • [7] Talking Face Generation with Expression-Tailored Generative Adversarial Network
    Zeng, Dan
    Liu, Han
    Lin, Hui
    Ge, Shiming
    MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, : 1716 - 1724
  • [8] Recognition-Aware HRRP Generation With Generative Adversarial Network
    Huang, Yue
    Wen, Yi
    Shi, Liangchao
    Ding, Xinghao
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2022, 19
  • [9] DAnet: DEPTH-AWARE NETWORK FOR CROWD COUNTING
    Van-Su Huynh
    Hoang Tran
    Huang, Ching-Chun
    2019 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2019, : 3001 - 3005
  • [10] Dast-Net: Depth-Aware Spatio-Temporal Network for Video Deblurring
    Zhu, Qi
    Xiao, Zeyu
    Huang, Jie
    Zhao, Feng
    Proceedings - IEEE International Conference on Multimedia and Expo, 2022, 2022-July