Single-Image 3D Human Pose and Shape Estimation Enhanced by Clothed 3D Human Reconstruction

被引:0
|
作者
Liu, Leyuan [1 ,2 ]
Gao, Yunqi [1 ]
Sun, Jianchi [1 ]
Chen, Jingying [1 ,2 ]
机构
[1] Cent China Normal Univ, Natl Engn Res Ctr E Learning, Wuhan, Peoples R China
[2] Cent China Normal Univ, Natl Engn Lab Educ Big Data, Wuhan, Peoples R China
基金
中国国家自然科学基金;
关键词
3D Human Pose and Shape Estimation; Clothed 3D Human Reconstruction; Graph Convolutional Network; SMPL Parameter Regression;
D O I
10.1007/978-981-99-9109-9_4
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
3D human pose and shape estimation and clothed 3D human reconstruction are two hot topics in the community of computer vision. 3D human pose and shape estimation aims to estimate the 3D poses and body shapes of "naked" humans under clothes, while clothed 3D human reconstruction refers to reconstructing the surfaces of humans wearing clothes. These two topics are closely related, but researchers usually study them separately. In this paper, we enhance the accuracy of the 3D human pose and body shape estimation by the reconstructed clothed 3D human models. Our method consists of two main components: the 3D body mesh recovery module and the clothed 3D human reconstruction module. In the 3D body mesh recovery module, an intermediate 3D body mesh is first recovered from the input image by a graph convolutional network (GCN), and then the 3D body pose and shape parameters are estimated by a regressor. In the clothed human reconstruction module, two clothed human surface models are respectively reconstructed under the guidance of the recovered 3D body mesh and the ground-truth 3D body mesh. At the training phase, losses which are described by the residuals among the two reconstructed clothed human models and ground truth are passed back into the 3D body mesh recovery module and used for boosting the body mesh recovery module. The quantitative and qualitative experimental results on THuman2.0, and LSP show that our method outperforms the current state-of-the-art 3D human pose and shape estimation methods.
引用
收藏
页码:33 / 44
页数:12
相关论文
共 50 条
  • [41] DeepHuman: 3D Human Reconstruction From a Single Image
    Zheng, Zerong
    Yu, Tao
    Wei, Yixuan
    Dai, Qionghai
    Liu, Yebin
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 7738 - 7748
  • [42] 3D Context-Aware PIFu for Clothed Human Reconstruction
    Liao, Tingting
    Zhu, Xiangyu
    Lei, Zhen
    Li, Stan Z.
    BIOMETRIC RECOGNITION (CCBR 2021), 2021, 12878 : 128 - 138
  • [43] Occlusion Resilient 3D Human Pose Estimation
    Roy, Soumava Kumar
    Badanin, Ilia
    Honari, Sina
    Fua, Pascal
    2024 INTERNATIONAL CONFERENCE IN 3D VISION, 3DV 2024, 2024, : 1198 - 1207
  • [44] A survey on monocular 3D human pose estimation
    Ji X.
    Fang Q.
    Dong J.
    Shuai Q.
    Jiang W.
    Zhou X.
    Virtual Reality and Intelligent Hardware, 2020, 2 (06): : 471 - 500
  • [45] Precise 3D Pose Estimation of Human Faces
    Pernek, Akos
    Hajder, Levente
    PROCEEDINGS OF THE 2014 9TH INTERNATIONAL CONFERENCE ON COMPUTER VISION, THEORY AND APPLICATIONS (VISAPP 2014), VOL 3, 2014, : 618 - 625
  • [46] A survey on deep 3D human pose estimation
    Neupane, Rama Bastola
    Li, Kan
    Boka, Tesfaye Fenta
    ARTIFICIAL INTELLIGENCE REVIEW, 2024, 58 (01)
  • [47] Deep 3D human pose estimation: A review
    Wang, Jinbao
    Tan, Shujie
    Zhen, Xiantong
    Xu, Shuo
    Zheng, Feng
    He, Zhenyu
    Shao, Ling
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2021, 210
  • [48] 3D Human Pose Estimation With Adversarial Learning
    Meng, Wenming
    Hu, Tao
    Shuai, Li
    2019 INTERNATIONAL CONFERENCE ON VIRTUAL REALITY AND VISUALIZATION (ICVRV), 2019, : 93 - 99
  • [49] Generative estimation of 3D human pose using shape contexts matching
    Zhao, Xu
    Liu, Yuncai
    COMPUTER VISION - ACCV 2007, PT I, PROCEEDINGS, 2007, 4843 : 419 - 429
  • [50] 3D human pose estimation by depth map
    Wu, Jianzhai
    Hu, Dewen
    Xiang, Fengtao
    Yuan, Xingsheng
    Su, Jiongming
    VISUAL COMPUTER, 2020, 36 (07): : 1401 - 1410