METACAP: Meta-learning Priors from Multi-view Imagery for Sparse-View Human Performance Capture and Rendering

被引:0
|
作者
Sun, Guoxing [1 ]
Dabral, Rishabh [1 ]
Fua, Pascal [2 ]
Theobalt, Christian [1 ]
Habermann, Marc [1 ]
机构
[1] Max Planck Inst Informat, Saarland Informat Campus, Saarbrucken, Germany
[2] Ecole Polytech Fed Lausanne, Lausanne, Switzerland
来源
关键词
Human Performance Capture; Meta Learning; EFFICIENT;
D O I
10.1007/978-3-031-72952-2_20
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Faithful human performance capture and free-view rendering from sparse RGB observations is a long-standing problem in Vision and Graphics. The main challenges are the lack of observations and the inherent ambiguities of the setting, e.g. occlusions and depth ambiguity. As a result, radiance fields, which have shown great promise in capturing high-frequency appearance and geometry details in dense setups, perform poorly when naively supervising them on sparse camera views, as the field simply overfits to the sparse-view inputs. To address this, we propose METACAP, a method for efficient and high-quality geometry recovery and novel view synthesis given very sparse or even a single view of the human. Our key idea is to meta-learn the radiance field weights solely from potentially sparse multi-view videos, which can serve as a prior when fine-tuning them on sparse imagery depicting the human. This prior provides a good network weight initialization, thereby effectively addressing ambiguities in sparse-view capture. Due to the articulated structure of the human body and motion-induced surface deformations, learning such a prior is non-trivial. Therefore, we propose to meta-learn the field weights in a pose-canonicalized space, which reduces the spatial feature range and makes feature learning more effective. Consequently, one can fine-tune our field parameters to quickly generalize to unseen poses, novel illumination conditions as well as novel and sparse (even monocular) camera views. For evaluating our method under different scenarios, we collect a new dataset, WILDDYNACAP, which contains subjects captured in, both, a dense camera dome and in-the-wild sparse camera rigs, and demonstrate superior results compared to recent state-of-the-art methods on, both, public and WILDDYNACAP dataset.
引用
收藏
页码:341 / 361
页数:21
相关论文
共 50 条
  • [41] NeuS: Learning Neural Implicit Surfaces by Volume Rendering for Multi-view Reconstruction
    Wang, Peng
    Liu, Lingjie
    Liu, Yuan
    Theobalt, Christian
    Komura, Taku
    Wang, Wenping
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [42] Incremental multi-view spectral clustering with sparse and connected graph learning
    Yin, Hongwei
    Hu, Wenjun
    Zhang, Zhao
    Lou, Jungang
    Miao, Minmin
    NEURAL NETWORKS, 2021, 144 : 260 - 270
  • [43] Lightweight Multi-person Total Motion Capture Using Sparse Multi-view Cameras
    Zhang, Yuxiang
    Li, Zhe
    An, Liang
    Li, Mengcheng
    Yu, Tao
    Liu, Yebin
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 5540 - 5549
  • [44] Performance analysis of a parallel multi-view rendering architecture using light fields
    Lages, Wallace
    Cordeiro, Carlucio
    Guedes, Dorgival
    VISUAL COMPUTER, 2009, 25 (10): : 947 - 958
  • [45] 3D Clothed Human Reconstruction from Sparse Multi-View Images
    Hong, Jin Gyu
    Noh, Seung Young
    Lee, Hee Kyung
    Cheong, Won Sik
    Chang, Ju Yong
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW, 2024, : 677 - 687
  • [46] Performance analysis of a parallel multi-view rendering architecture using light fields
    Wallace Lages
    Carlúcio Cordeiro
    Dorgival Guedes
    The Visual Computer, 2009, 25 : 947 - 958
  • [47] A novel multi-view learning developed from single-view patterns
    Wang, Zhe
    Chen, Songcan
    Gao, Daqi
    PATTERN RECOGNITION, 2011, 44 (10-11) : 2395 - 2413
  • [48] Multiplicative Sparse Feature Decomposition for Efficient Multi-View Multi-Task Learning
    Sun, Lu
    Nguyen, Canh Hao
    Mamitsuka, Hiroshi
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 3506 - 3512
  • [49] Improving Neural Surface Reconstruction with Feature Priors from Multi-view Images
    Ren, Xinlin
    Cao, Chenjie
    Fu, Yanwei
    Xu, Xiangyang
    COMPUTER VISION - ECCV 2024, PT LVIII, 2025, 15116 : 445 - 463
  • [50] MVDiffusion++: A Dense High-resolution Multi-view Diffusion Model for Single or Sparse-view 3D Object Reconstruction
    Tang, Shitao
    Chen, Jiacheng
    Wang, Dilin
    Tang, Chengzhou
    Zhang, Fuyang
    Fan, Yuchen
    Chandra, Vikas
    Furukawa, Yasutaka
    Ranjan, Rakesh
    arXiv,