Learning to Estimate Pose and Shape of Hand-Held Objects from RGB Images

被引:0
|
作者
Kokic, Mia [1 ]
Kragic, Danica [1 ]
Bohg, Jeannette [2 ]
机构
[1] KTH, EECS, Div Robot Percept & Learning, Stockholm, Sweden
[2] Stanford Univ, Dept Comp Sci, Stanford, CA 94305 USA
基金
瑞典研究理事会;
关键词
GRASP;
D O I
10.1109/iros40897.2019.8967961
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We develop a system for modeling hand-object interactions in 3D from RGB images that show a hand which is holding a novel object from a known category. We design a Convolutional Neural Network (CNN) for Hand-held Object Pose and Shape estimation called HOPS-Net and utilize prior work to estimate the hand pose and configuration. We leverage the insight that information about the hand facilitates object pose and shape estimation by incorporating the hand into both training and inference of the object pose and shape as well as the refinement of the estimated pose. The network is trained on a large synthetic dataset of objects in interaction with a human hand. To bridge the gap between real and synthetic images, we employ an image-to-image translation model (Augmented CycleGAN) that generates realistically textured objects given a synthetic rendering. This provides a scalable way of generating annotated data for training HOPS-Net. Our quantitative experiments show that even noisy hand parameters significantly help object pose and shape estimation. The qualitative experiments show results of pose and shape estimation of objects held by a hand "in the wild".
引用
收藏
页码:3980 / 3987
页数:8
相关论文
共 50 条
  • [1] Learning to Estimate 3D Hand Pose from Single RGB Images
    Zimmermann, Christian
    Brox, Thomas
    2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 4913 - 4921
  • [2] FreiHAND: A Dataset for Markerless Capture of Hand Pose and Shape from Single RGB Images
    Zimmermann, Christian
    Ceylan, Duygu
    Yang, Jimei
    Russell, Bryan
    Argus, Max
    Brox, Thomas
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 813 - 822
  • [3] Hand Pose Estimation from RGB Images Based on Deep Learning: A Survey
    Liu, Yang
    Jiang, Jie
    Sun, Jiahao
    2021 IEEE 7TH INTERNATIONAL CONFERENCE ON VIRTUAL REALITY (ICVR 2021), 2021, : 82 - 89
  • [4] Reconstructing Hand-Held Objects from Monocular Video
    Huang, Di
    Ji, Xiaopeng
    He, Xingyi
    Sun, Jiaming
    He, Tong
    Shuai, Qing
    Ouyang, Wanli
    Zhou, Xiaowei
    PROCEEDINGS SIGGRAPH ASIA 2022, 2022,
  • [5] A Shared Pose Regression Network for Pose Estimation of Objects from RGB Images
    Bengtson, Stefan Hein
    Astrom, Hampus
    Moeslund, Thomas B.
    Topp, Elin A.
    Krueger, Volker
    2022 16TH INTERNATIONAL CONFERENCE ON SIGNAL-IMAGE TECHNOLOGY & INTERNET-BASED SYSTEMS, SITIS, 2022, : 91 - 97
  • [6] RGB-D Object Recognition from Hand-Held Object Teaching
    Qiao, Leixian
    Li, Xue
    Jiang, Shuqiang
    8TH INTERNATIONAL CONFERENCE ON INTERNET MULTIMEDIA COMPUTING AND SERVICE (ICIMCS2016), 2016, : 31 - 34
  • [7] 3D hand pose and shape estimation from RGB images for keypoint-based hand gesture recognition
    Avola, Danilo
    Cinque, Luigi
    Fagioli, Alessio
    Foresti, Gian Luca
    Fragomeni, Adriano
    Pannone, Daniele
    PATTERN RECOGNITION, 2022, 129
  • [8] HAND-HELD SPECTRAL RADIOMETER TO ESTIMATE GRAMINEOUS BIOMASS
    PEARSON, RL
    MILLER, LD
    TUCKER, CJ
    APPLIED OPTICS, 1976, 15 (02): : 416 - 418
  • [9] HAPTICALLY PERCEIVING THE DISTANCES REACHABLE WITH HAND-HELD OBJECTS
    SOLOMON, HY
    TURVEY, MT
    JOURNAL OF EXPERIMENTAL PSYCHOLOGY-HUMAN PERCEPTION AND PERFORMANCE, 1988, 14 (03) : 404 - 427
  • [10] Fitts's law for evaluating the recovery from movements of hand-held objects
    State Key Laboratory of Tribology, Department of Precision, Tsinghua University, Beijing 100084, China
    Qinghua Daxue Xuebao, 2006, 11 (1848-1850):