Visual-inertial object tracking: Incorporating camera pose into motion models

被引:3
|
作者
Shahbazi, Mohammad [1 ]
Mirtajadini, Seyed Hojat [2 ]
Fahimi, Hamidreza [3 ]
机构
[1] Iran Univ Sci & Technol, Sch Mech Engn, Hengam St, Tehran 1684613114, Iran
[2] Univ Tehran, Fac New Sci & Technol, North Kargar St, Tehran 1439957131, Iran
[3] Amirkabir Univ Technol, Dept Aerosp Engn, Hafez Ave, Tehran 1591634311, Iran
关键词
Visual object tracking; Object tracking dataset; Aerial robot; Deep learning; Visual-inertial navigation;
D O I
10.1016/j.eswa.2023.120483
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Visual object tracking for autonomy of aerial robots could become challenging especially in the presence of target or camera fast motions and long-term occlusions. This paper presents a visual-inertial tracking paradigm by incorporating the camera kinematics states into the visual object tracking pipelines. We gathered a dataset of image sequences with the addition of camera's position and orientation measurements as well as the object's position measurement. For the cases of long-term object occlusion, we provide ground-truth boxes derived from mapping the measured object position onto the image frame. A search zone proposal method is developed based on the estimation of object future position represented in the inertial frame and projected back onto the image frame using the camera states. This search zone, which is robust to fast camera/target motions, is fused into the original search zone settings of the base tracker. Also proposed is a measure indicating the confidence of a tracking structure in keeping track of a correct target and reporting the tracking failure in-time. Accordingly, the model updating mechanism of base tracker is modulated to avoid recovering of wrong objects as the target. The proposed modifications are benchmarked on nine visual object tracking algorithms including five state-of-art deep learning structures, namely DiMP, PrDiMP, KYS, ToMP, and MixFormer. Most of the trackers are remarkably improved by the modifications with up to 8% increase in precision. Modified PrDiMP tracker yields the best precision of 68.4%, more than all considered original (and modified) trackers. Source code and dataset are made available online.1
引用
收藏
页数:15
相关论文
共 50 条
  • [31] Visual-Inertial Odometry With Online Calibration of Velocity-Control Based Kinematic Motion Models
    Li, Haolong
    Stueckler, Joerg
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (03) : 6415 - 6422
  • [32] Relative motion estimation using visual-inertial optical flow
    He, Hongsheng
    Li, Yan
    Tan, Jindong
    AUTONOMOUS ROBOTS, 2018, 42 (03) : 615 - 629
  • [33] A fast initialization method of Visual-Inertial Odometry based on monocular camera
    Huang, Lixiao
    Pan, Shuguo
    Wang, Shuai
    Zeng, Pan
    Ye, Fei
    PROCEEDINGS OF 5TH IEEE CONFERENCE ON UBIQUITOUS POSITIONING, INDOOR NAVIGATION AND LOCATION-BASED SERVICES (UPINLBS), 2018, : 70 - 74
  • [34] Minimalistic sensor design in visual-inertial structure from motion
    Martinelli, Agostino
    2015 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2015, : 3313 - 3318
  • [35] Visual-Inertial Ego-Motion Estimation for Humanoid Platforms
    Tsotsos, Konstantine
    Pretto, Alberto
    Soatto, Stefano
    2012 12TH IEEE-RAS INTERNATIONAL CONFERENCE ON HUMANOID ROBOTS (HUMANOIDS), 2012, : 704 - 711
  • [36] VI-RPE: Visual-Inertial Relative Pose Estimation for Aerial Vehicles
    Teixeira, Lucas
    Maffra, Fabiola
    Moos, Marco
    Chli, Margarita
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2018, 3 (04): : 2770 - 2777
  • [37] Rapid Initialization using Relative Pose Constraints in Stereo Visual-Inertial Odometry
    Jung, Jae Hyung
    Chung, Jae Young
    Cha, Jaehyuck
    Park, Chan Gook
    2019 IEEE 15TH INTERNATIONAL CONFERENCE ON CONTROL AND AUTOMATION (ICCA), 2019, : 969 - 974
  • [38] Scale-Aware Monocular Visual-Inertial Pose Estimation for Aerial Robots
    Xiong, Minjun
    Lu, Huimin
    Xiong, Dan
    Xiao, Junhao
    Lv, Ming
    2017 CHINESE AUTOMATION CONGRESS (CAC), 2017, : 7030 - 7034
  • [39] A Framework for Visual-Inertial Object-Level Simultaneous Localization and Mapping
    Jung, Jae Hyung
    Park, Chan Gook
    2023 IEEE/ION POSITION, LOCATION AND NAVIGATION SYMPOSIUM, PLANS, 2023, : 1335 - 1340
  • [40] VIFTrack!-visual-inertial feature tracking based on affine photometric warping
    Aufderheide, Dominik
    Krybus, Werner
    Edwards, Gerard
    COMPUTATIONAL VISION AND MEDICAL IMAGE PROCESSING IV, 2014, : 155 - 160