Keypoint-Based Category-Level Object Pose Tracking from an RGB Sequence with Uncertainty Estimation

被引:0
|
作者
Lin, Yunzhi [1 ,2 ]
Tremblay, Jonathan [1 ]
Tyree, Stephen [1 ]
Vela, Patricio A. [2 ]
Birchfield, Stan [1 ]
机构
[1] NVIDIA, Santa Clara, CA 95051 USA
[2] Georgia Inst Technol, Atlanta, GA 30332 USA
关键词
D O I
10.1109/ICRA.46639.2022.9811720
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We propose a single-stage, category-level 6-DoF pose estimation algorithm that simultaneously detects and tracks instances of objects within a known category. Our method takes as input the previous and current frame from a monocular RGB video, as well as predictions from the previous frame, to predict the bounding cuboid and 6-DoF pose (up to scale). Internally, a deep network predicts distributions over object keypoints (vertices of the bounding cuboid) in image coordinates, after which a novel probabilistic filtering process integrates across estimates before computing the final pose using PnP. Our framework allows the system to take previous uncertainties into consideration when predicting the current frame, resulting in predictions that are more accurate and stable than single frame methods. Extensive experiments show that our method outperforms existing approaches on the challenging Objectron benchmark of annotated object videos. We also demonstrate the usability of our work in an augmented reality setting.
引用
收藏
页数:7
相关论文
共 50 条
  • [21] GS-Pose: Category-Level Object Pose Estimation via Geometric and Semantic Correspondence
    Wang, Pengyuan
    Ikeda, Takuya
    Lee, Robert
    Nishiwaki, Koichi
    COMPUTER VISION - ECCV 2024, PT XXVII, 2025, 15085 : 108 - 126
  • [22] HS-Pose: Hybrid Scope Feature Extraction for Category-level Object Pose Estimation
    Zheng, Linfang
    Wang, Chen
    Sun, Yinghan
    Dasgupta, Esha
    Chen, Hua
    Leonardis, Ales
    Zhang, Wei
    Chang, Hyung Jin
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 17163 - 17173
  • [23] Instance-Adaptive and Geometric-Aware Keypoint Learning for Category-Level 6D Object Pose Estimation
    Lin, Xiao
    Yang, Wenfei
    Gao, Yuan
    Zhan, Tianzhu
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 21040 - 21049
  • [24] An efficient network for category-level 6D object pose estimation
    Sun, Shantong
    Liu, Rongke
    Sun, Shuqiao
    Yang, Xinxin
    Lu, Guangshan
    SIGNAL IMAGE AND VIDEO PROCESSING, 2021, 15 (07) : 1643 - 1651
  • [25] CatFormer: Category-Level 6D Object Pose Estimation with Transformer
    Yu, Sheng
    Zhai, Di-Hua
    Xia, Yuanqing
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 7, 2024, : 6808 - 6816
  • [26] RANSAC Optimization for Category-level 6D Object Pose Estimation
    Chen, Ying
    Kang, Guixia
    Wang, Yiping
    2020 5TH INTERNATIONAL CONFERENCE ON MECHANICAL, CONTROL AND COMPUTER ENGINEERING (ICMCCE 2020), 2020, : 50 - 56
  • [27] Keypoint-based contextual representations for hand pose estimation
    Li, Weiwei
    Du, Rong
    Chen, Shudong
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (10) : 28357 - 28372
  • [28] An efficient network for category-level 6D object pose estimation
    Shantong Sun
    Rongke Liu
    Shuqiao Sun
    Xinxin Yang
    Guangshan Lu
    Signal, Image and Video Processing, 2021, 15 : 1643 - 1651
  • [29] Keypoint-based contextual representations for hand pose estimation
    Weiwei Li
    Rong Du
    Shudong Chen
    Multimedia Tools and Applications, 2024, 83 : 28357 - 28372
  • [30] Attention-guided RGB-D Fusion Network for Category-level 6D Object Pose Estimation
    Wang, Hao
    Li, Weiming
    Kim, Jiyeon
    Wang, Qiang
    2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 10651 - 10658