Visual-inertial object tracking: Incorporating camera pose into motion models

被引:3
|
作者
Shahbazi, Mohammad [1 ]
Mirtajadini, Seyed Hojat [2 ]
Fahimi, Hamidreza [3 ]
机构
[1] Iran Univ Sci & Technol, Sch Mech Engn, Hengam St, Tehran 1684613114, Iran
[2] Univ Tehran, Fac New Sci & Technol, North Kargar St, Tehran 1439957131, Iran
[3] Amirkabir Univ Technol, Dept Aerosp Engn, Hafez Ave, Tehran 1591634311, Iran
关键词
Visual object tracking; Object tracking dataset; Aerial robot; Deep learning; Visual-inertial navigation;
D O I
10.1016/j.eswa.2023.120483
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Visual object tracking for autonomy of aerial robots could become challenging especially in the presence of target or camera fast motions and long-term occlusions. This paper presents a visual-inertial tracking paradigm by incorporating the camera kinematics states into the visual object tracking pipelines. We gathered a dataset of image sequences with the addition of camera's position and orientation measurements as well as the object's position measurement. For the cases of long-term object occlusion, we provide ground-truth boxes derived from mapping the measured object position onto the image frame. A search zone proposal method is developed based on the estimation of object future position represented in the inertial frame and projected back onto the image frame using the camera states. This search zone, which is robust to fast camera/target motions, is fused into the original search zone settings of the base tracker. Also proposed is a measure indicating the confidence of a tracking structure in keeping track of a correct target and reporting the tracking failure in-time. Accordingly, the model updating mechanism of base tracker is modulated to avoid recovering of wrong objects as the target. The proposed modifications are benchmarked on nine visual object tracking algorithms including five state-of-art deep learning structures, namely DiMP, PrDiMP, KYS, ToMP, and MixFormer. Most of the trackers are remarkably improved by the modifications with up to 8% increase in precision. Modified PrDiMP tracker yields the best precision of 68.4%, more than all considered original (and modified) trackers. Source code and dataset are made available online.1
引用
收藏
页数:15
相关论文
共 50 条
  • [21] Tracking 3-D Motion of Dynamic Objects Using Monocular Visual-Inertial Sensing
    Qiu, Kejie
    Qin, Tong
    Gao, Wenliang
    Shen, Shaojie
    IEEE TRANSACTIONS ON ROBOTICS, 2019, 35 (04) : 799 - 816
  • [22] Covariance Estimation for Pose Graph Optimization in Visual-Inertial Navigation Systems
    Shi, Pengcheng
    Zhu, Zhikai
    Sun, Shiying
    Rong, Zheng
    Zhao, Xiaoguang
    Tan, Min
    IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2023, 8 (06): : 3657 - 3667
  • [23] Dataset: HoloSet - A Dataset for Visual-Inertial Pose Estimation in Extended Reality
    Chandio, Yasra
    Bashir, Noman
    Anwar, Fatima M.
    PROCEEDINGS OF THE TWENTIETH ACM CONFERENCE ON EMBEDDED NETWORKED SENSOR SYSTEMS, SENSYS 2022, 2022, : 1014 - 1019
  • [24] Fusion of Monocular Visual-Inertial Measurements for Three Dimensional Pose Estimatione
    Perez-Paina, Gonzalo
    Paz, Claudio
    Kulich, Miroslav
    Saska, Martin
    Araguas, Gaston
    MODELLING AND SIMULATION FOR AUTONOMOUS SYSTEMS, MESAS 2016, 2016, 9991 : 242 - 260
  • [25] Wearable Heading Estimation for Motion Tracking in Health Care by Adaptive Fusion of Visual-Inertial Measurements
    Zhang, Yinlong
    Liang, Wei
    He, Hongsheng
    Tan, Jindong
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2018, 22 (06) : 1732 - 1743
  • [26] Visual-Inertial Fusion-Based Human Pose Estimation: A Review
    Li, Tong
    Yu, Haoyong
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2023, 72
  • [27] Hierarchical Quadtree Feature Optical Flow Tracking Based Sparse Pose-Graph Visual-Inertial SLAM
    Xie, Hongle
    Chen, Weidong
    Wang, Jingchuan
    Wang, Hesheng
    2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 58 - 64
  • [28] Multi-Robot Joint Visual-Inertial Localization and 3-D Moving Object Tracking
    Zhu, Pengxiang
    Ren, Wei
    2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 11573 - 11580
  • [29] A Moving Target Tracking System of Quadrotors with Visual-Inertial Localization
    Lin, Ziyue
    Xu, Wenbo
    Wang, Wei
    2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA, 2023, : 3296 - 3302
  • [30] USING OPTICAL FLOW FOR FILLING THE GAPS IN VISUAL-INERTIAL TRACKING
    Bleser, Gabriele
    Hendeby, Gustaf
    18TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO-2010), 2010, : 1836 - 1840