Improving 3D Human Pose Estimation via 3D Part Affinity Fields

被引:9
|
作者
Liu, Ding [1 ]
Zhao, Zixu [1 ]
Wang, Xinchao [2 ]
Hu, Yuxiao [3 ]
Zhang, Lei [4 ]
Huang, Thomas S. [1 ]
机构
[1] Univ Illinois, Urbana, IL 61801 USA
[2] Stevens Inst Technol, Hoboken, NJ 07030 USA
[3] Huawei Technol Inc USA, Santa Clara, CA USA
[4] Microsoft, Bellevue, WA USA
来源
2019 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV) | 2019年
关键词
REPRESENTATION;
D O I
10.1109/WACV.2019.00112
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
3D human pose estimation from monocular images has become a heated area in computer vision recently. For years, most deep neural network based practices have adopted either an end-to-end approach, or a two-stage approach. An end-to-end network typically estimates 3D human poses directly from 2D input images, but it suffers from the shortage of 3D human pose data. It is also obscure to know if the inaccuracy stems from limited visual understanding or 2D-to-3D mapping. Whereas a two-stage directly lifts those 2D keypoint outputs to the 3D space, after utilizing an existing network for 2D keypoint detections. However, they tend to ignore some useful contextual hints from the 2D raw image pixels. In this paper, we introduce a two-stage architecture that can eliminate the main disadvantages of both these approaches. During the first stage we use an existing stateof- the-art detector to estimate 2D poses. To add more contextual information to help lifting 2D poses to 3D poses, we propose 3D Part Affinity Fields (3D-PAFs). We use 3D-PAFs to infer 3D limb vectors, and combine them with 2D poses to regress the 3D coordinates. We trained and tested our proposed framework on Human3.6M, the most popular 3D human pose benchmark dataset. Our approach achieves the state-of-the-art performance, which proves that with right selections of contextual information, a simple regression model can be very powerful in estimating 3D poses.
引用
收藏
页码:1004 / 1013
页数:10
相关论文
共 50 条
  • [21] 3D human pose estimation by depth map
    Wu, Jianzhai
    Hu, Dewen
    Xiang, Fengtao
    Yuan, Xingsheng
    Su, Jiongming
    VISUAL COMPUTER, 2020, 36 (07): : 1401 - 1410
  • [22] 3D Human Pose Estimation using 2D Body Part Detectors
    Barbulescu, Adela
    Gong, Wenjuan
    Gonzalez, Jordi
    Moeslund, Thomas B.
    Xavier Roca, F.
    2012 21ST INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR 2012), 2012, : 2484 - 2487
  • [23] MONOCULAR 3D HUMAN POSE ESTIMATION BY CLASSIFICATION
    Greif, Thomas
    Lienhart, Rainer
    Sengupta, Debabrata
    2011 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2011,
  • [24] 3D human pose estimation by depth map
    Jianzhai Wu
    Dewen Hu
    Fengtao Xiang
    Xingsheng Yuan
    Jiongming Su
    The Visual Computer, 2020, 36 : 1401 - 1410
  • [25] 3D Human Pose Estimation=2D Pose Estimation plus Matching
    Chen, Ching-Hang
    Ramanan, Deva
    30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 5759 - 5767
  • [26] Stabilization of 3D pose estimation
    Neddermeyer, W
    Schnell, M
    Winkler, W
    Lilienthal, A
    APPLICATIONS OF GEOMETRIC ALGEBRA IN COMPUTER SCIENCE AND ENGINEERING, 2002, : 385 - 394
  • [27] Learning Dynamical Human-Joint Affinity for 3D Pose Estimation in Videos
    Zhang, Junhao
    Wang, Yali
    Zhou, Zhipeng
    Luan, Tianyu
    Wang, Zhe
    Qiao, Yu
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 7914 - 7925
  • [28] Accurate 3D Hand Pose Estimation for Whole-Body 3D Human Mesh Estimation
    Moon, Gyeongsik
    Choi, Hongsuk
    Lee, Kyoung Mu
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2022, 2022, : 2307 - 2316
  • [29] Pointless Pose: Part Affinity Field-Based 3D Pose Estimation without Detecting Keypoints
    Wang, Jue
    Luo, Zhigang
    ELECTRONICS, 2021, 10 (08)
  • [30] Efficient Human Pose Estimation via 3D Event Point Cloud
    Chen, Jiaan
    Shi, Hao
    Ye, Yaozu
    Yang, Kailun
    Sun, Lei
    Wang, Kaiwei
    2022 INTERNATIONAL CONFERENCE ON 3D VISION, 3DV, 2022, : 104 - 113