Dual-view 3D human pose estimation without camera parameters for action recognition

被引:8
|
作者
Liu, Long [1 ]
Yang, Le [1 ]
Chen, Wanjun [2 ]
Gao, Xin [1 ]
机构
[1] Xian Univ Technol, Sch Automat & Informat Engn, 5 Jinhua South Rd, Xian, Shaanxi, Peoples R China
[2] Xian Univ Technol, Dept Informat Sci, Xian, Shaanxi, Peoples R China
基金
中国国家自然科学基金;
关键词
Cameras - Virtual reality;
D O I
10.1049/ipr2.12277
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The purpose of 3D human pose estimation is to estimate the 3D coordinates of key points of the human body directly from images. Although multi-view based methods have better performance and higher precision of coordinate estimation than a single-view based, they need to know the camera parameters. In order to effectively avoid the restriction of this constraint and improve the generalizability of the model, a dual-view single-person 3D pose estimation method without camera parameters is proposed. This method first uses the 2D pose estimation network HR-net to estimate the 2D joint point coordinates from two images with different views, and then inputs them into the 3D regression network to generate the final 3D joint point coordinates. In order to make the 3D regression network fully learn the spatial structure relationship of the human body and the transformation projection relationship between different views, a self-supervised training method is designed based on a 3D human pose orthogonal projection model to generate the virtual views. In the pose estimation experiments on the Human3.6 dataset, this method achieves a significantly improved estimation error of 34.5 mm. Furthermore, an action recognition based on the human poses extracted by the proposed method is conducted, and an accuracy of 83.19% is obtained.
引用
收藏
页码:3433 / 3440
页数:8
相关论文
共 50 条
  • [1] Distributed RGBD Camera Network for 3D Human Pose Estimation and Action Recognition
    Li, Junwei
    Liu, Guoliang
    Tian, Guohui
    Zhu, Xianglai
    Wang, Ziren
    2018 21ST INTERNATIONAL CONFERENCE ON INFORMATION FUSION (FUSION), 2018, : 1554 - 1558
  • [2] Multi-person 3D pose estimation from multi-view without extrinsic camera parameters
    Xu, Daoliang
    Zheng, Tianyou
    Zhang, Yang
    Yang, Xiaodong
    Fu, Weiwei
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 266
  • [3] Dual-view 3D object recognition and detection via Lidar point cloud and camera image
    Li, Jing
    Li, Rui
    Li, Jiehao
    Wang, Junzheng
    Wu, Qingbin
    Liu, Xu
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2022, 150
  • [4] 2D Action Recognition Serves 3D Human Pose Estimation
    Gall, Juergen
    Yao, Angela
    Van Gool, Luc
    COMPUTER VISION-ECCV 2010, PT III, 2010, 6313 : 425 - 438
  • [5] Joint Camera Pose Estimation and 3D Human Pose Estimation in a Multi-camera Setup
    Puwein, Jens
    Ballan, Luca
    Ziegler, Remo
    Pollefeys, Marc
    COMPUTER VISION - ACCV 2014, PT II, 2015, 9004 : 473 - 487
  • [6] View Invariant 3D Human Pose Estimation
    Wei, Guoqiang
    Lan, Cuiling
    Zeng, Wenjun
    Chen, Zhibo
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2020, 30 (12) : 4601 - 4610
  • [7] Snapshot dual-view 3D imaging
    Ji, Chao
    Fang, Mengyan
    Xin, Liwei
    He, Kai
    Li, Yahui
    Wang, Xing
    Tian, Jinshou
    AIP ADVANCES, 2023, 13 (04)
  • [8] Temporal 3D Human Pose Estimation for Action Recognition from Arbitrary Viewpoints
    Musallam, Mohamed Adel
    Baptista, Renato
    Al Ismaeil, Kassem
    Aouada, Djamila
    2019 6TH INTERNATIONAL CONFERENCE ON COMPUTATIONAL SCIENCE AND COMPUTATIONAL INTELLIGENCE (CSCI 2019), 2019, : 253 - 258
  • [9] Cross View Fusion for 3D Human Pose Estimation
    Qiu, Haibo
    Wang, Chunyu
    Wang, Jingdong
    Wang, Naiyan
    Zeng, Wenjun
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 4341 - 4350
  • [10] VIEW-INVARIANT ACTION RECOGNITION FROM RGB DATA VIA 3D POSE ESTIMATION
    Baptista, Renato
    Ghorbel, Enjie
    Papadopoulos, Konstantinos
    Demisse, Girum G.
    Aouada, Djamila
    Ottersten, Bjorn
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 2542 - 2546