Self-supervised Neural Articulated Shape and Appearance Models

被引:10
|
作者
Wei, Fangyin [1 ,2 ]
Chabra, Rohan [2 ]
Ma, Lingni [2 ]
Lassner, Christoph [2 ]
Zollhoefer, Michael [2 ]
Rusinkiewicz, Szymon [1 ]
Sweeney, Chris [2 ]
Newcombe, Richard [2 ]
Slavcheva, Mira [2 ]
机构
[1] Princeton Univ, Princeton, NJ 08544 USA
[2] Real Labs Res, Menlo Pk, CA USA
关键词
D O I
10.1109/CVPR52688.2022.01536
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Learning geometry, motion, and appearance priors of object classes is important for the solution of a large variety of computer vision problems. While the majority of approaches has focused on static objects, dynamic objects, especially with controllable articulation, are less explored. We propose a novel approach for learning a representation of the geometry, appearance, and motion of a class of articulated objects given only a set of color images as input. In a self-supervised manner, our novel representation learns shape, appearance, and articulation codes that enable independent control of these semantic dimensions. Our model is trained end-to-end without requiring any articulation annotations. Experiments show that our approach performs well for different joint types, such as revolute and prismatic joints, as well as different combinations of these joints. Compared to state of the art that uses direct 3D supervision and does not output appearance, we recover more faithful geometry and appearance from 2D observations only. In addition, our representation enables a large variety of applications, such as few-shot reconstruction, the generation of novel articulations, and novel view-synthesis. Project page: https://weify627.github.io/nasam/.
引用
收藏
页码:15795 / 15805
页数:11
相关论文
共 50 条
  • [31] Self-supervised role learning for graph neural networks
    Aravind Sankar
    Junting Wang
    Adit Krishnan
    Hari Sundaram
    Knowledge and Information Systems, 2022, 64 : 2091 - 2121
  • [32] Self-Supervised Online Learning of Appearance for 3D Tracking
    Lee, Bhoram
    Lee, Daniel D.
    2017 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2017, : 4930 - 4937
  • [33] Self-Supervised Position Debiasing for Large Language Models
    Liu, Zhongkun
    Chen, Zheng
    Zhang, Mengqi
    Ren, Zhaochun
    Chen, Zhumin
    Ren, Pengjie
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 2897 - 2917
  • [34] ProsAudit, a prosodic benchmark for self-supervised speech models
    de Seyssel, Maureen
    Lavechin, Marvin
    Titeux, Hadrien
    Thomas, Arthur
    Virlet, Gwendal
    Revilla, Andrea Santos
    Wisniewski, Guillaume
    Ludusan, Bogdan
    Dupoux, Emmanuel
    INTERSPEECH 2023, 2023, : 2963 - 2967
  • [35] Fusion of Self-supervised Learned Models for MOS Prediction
    Yang, Zhengdong
    Zhou, Wangjin
    Chu, Chenhui
    Li, Sheng
    Dabre, Raj
    Rubino, Raphael
    Zhao, Yi
    INTERSPEECH 2022, 2022, : 5443 - 5447
  • [36] PHONEME SEGMENTATION USING SELF-SUPERVISED SPEECH MODELS
    Strgar, Luke
    Harwath, David
    2022 IEEE SPOKEN LANGUAGE TECHNOLOGY WORKSHOP, SLT, 2022, : 1067 - 1073
  • [37] Self-supervised MRI Reconstruction with Unrolled Diffusion Models
    Korkmaz, Yilmaz
    Cukur, Tolga
    Patel, Vishal M.
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2023, PT X, 2023, 14229 : 491 - 501
  • [38] How Well Do Self-Supervised Models Transfer?
    Ericsson, Linus
    Gouk, Henry
    Hospedales, Timothy M.
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 5410 - 5419
  • [39] The Efficacy of Self-Supervised Speech Models as Audio Representations
    Wu, Tung-Yu
    Hsu, Tsu-Yuan
    Li, Chen-An
    Lin, Tzu-Han
    Lee, Hung-yi
    HEAR: HOLISTIC EVALUATION OF AUDIO REPRESENTATIONS, VOL 166, 2021, 166 : 90 - 110
  • [40] Self-supervised Shape Completion via Involution and Implicit Correspondences
    Gao, Mengya
    Chhatkuli, Ajad
    Postels, Janis
    Van Gool, Luc
    Tombari, Federico
    COMPUTER VISION - ECCV 2024, PT LVIII, 2025, 15116 : 212 - 229