METACAP: Meta-learning Priors from Multi-view Imagery for Sparse-View Human Performance Capture and Rendering

被引:0
|
作者
Sun, Guoxing [1 ]
Dabral, Rishabh [1 ]
Fua, Pascal [2 ]
Theobalt, Christian [1 ]
Habermann, Marc [1 ]
机构
[1] Max Planck Inst Informat, Saarland Informat Campus, Saarbrucken, Germany
[2] Ecole Polytech Fed Lausanne, Lausanne, Switzerland
来源
关键词
Human Performance Capture; Meta Learning; EFFICIENT;
D O I
10.1007/978-3-031-72952-2_20
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Faithful human performance capture and free-view rendering from sparse RGB observations is a long-standing problem in Vision and Graphics. The main challenges are the lack of observations and the inherent ambiguities of the setting, e.g. occlusions and depth ambiguity. As a result, radiance fields, which have shown great promise in capturing high-frequency appearance and geometry details in dense setups, perform poorly when naively supervising them on sparse camera views, as the field simply overfits to the sparse-view inputs. To address this, we propose METACAP, a method for efficient and high-quality geometry recovery and novel view synthesis given very sparse or even a single view of the human. Our key idea is to meta-learn the radiance field weights solely from potentially sparse multi-view videos, which can serve as a prior when fine-tuning them on sparse imagery depicting the human. This prior provides a good network weight initialization, thereby effectively addressing ambiguities in sparse-view capture. Due to the articulated structure of the human body and motion-induced surface deformations, learning such a prior is non-trivial. Therefore, we propose to meta-learn the field weights in a pose-canonicalized space, which reduces the spatial feature range and makes feature learning more effective. Consequently, one can fine-tune our field parameters to quickly generalize to unseen poses, novel illumination conditions as well as novel and sparse (even monocular) camera views. For evaluating our method under different scenarios, we collect a new dataset, WILDDYNACAP, which contains subjects captured in, both, a dense camera dome and in-the-wild sparse camera rigs, and demonstrate superior results compared to recent state-of-the-art methods on, both, public and WILDDYNACAP dataset.
引用
收藏
页码:341 / 361
页数:21
相关论文
共 50 条
  • [21] Retargeted multi-view classification via structured sparse learning
    Wang, Zhi
    Shen, Zhencai
    Zou, Hui
    Zhong, Ping
    Chen, Yingyi
    SIGNAL PROCESSING, 2022, 197
  • [22] Adaptive sparse graph learning for multi-view spectral clustering
    Qingjiang Xiao
    Shiqiang Du
    Kaiwu Zhang
    Jinmei Song
    Yixuan Huang
    Applied Intelligence, 2023, 53 : 14855 - 14875
  • [23] Sparse multi-view learning and its application in disease diagnosis
    Wu Z.
    Tian Y.
    Fu S.
    Xitong Gongcheng Lilun yu Shijian/System Engineering Theory and Practice, 2023, 43 (02): : 556 - 567
  • [24] Multi-view spectral clustering via sparse graph learning
    Hu, Zhanxuan
    Nie, Feiping
    Chang, Wei
    Hao, Shuzheng
    Wang, Rong
    Li, Xuelong
    NEUROCOMPUTING, 2020, 384 : 1 - 10
  • [25] Sparse Graph Tensor Learning for Multi-View Spectral Clustering
    Chen, Man-Sheng
    Li, Zhi-Yuan
    Lin, Jia-Qi
    Wang, Chang-Dong
    Huang, Dong
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2024, : 1 - 10
  • [26] Sparse Graph Tensor Learning for Multi-View Spectral Clustering
    Chen, Man-Sheng
    Li, Zhi-Yuan
    Lin, Jia-Qi
    Wang, Chang-Dong
    Huang, Dong
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2024, 8 (05): : 3534 - 3543
  • [27] A weighted multi-view clustering via sparse graph learning
    Zhou, Jie
    Zhang, Runxin
    CLUSTER COMPUTING-THE JOURNAL OF NETWORKS SOFTWARE TOOLS AND APPLICATIONS, 2024, 27 (10): : 13517 - 13530
  • [28] Perceptually learning multi-view sparse representation for scene categorization
    Yin, Weibin
    Xu, Dongsheng
    Wang, Zheng
    Zhao, Zhijun
    Chen, Chao
    Yao, Yiyang
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2019, 60 : 59 - 63
  • [29] Performance of sparse-view CT reconstruction with multi-directional gradient operators
    Hsieh, Chia-Jui
    Jin, Shih-Chun
    Chen, Jyh-Cheng
    Kuo, Chih-Wei
    Wang, Ruei-Teng
    Chu, Woei-Chyn
    PLOS ONE, 2019, 14 (01):
  • [30] Robust Multi-view Representation: A Unified Perspective from Multi-view Learning to Domain Adaption
    Ding, Zhengming
    Shao, Ming
    Fu, Yun
    PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018, : 5434 - 5440