Visual-inertial object tracking: Incorporating camera pose into motion models

被引:3
|
作者
Shahbazi, Mohammad [1 ]
Mirtajadini, Seyed Hojat [2 ]
Fahimi, Hamidreza [3 ]
机构
[1] Iran Univ Sci & Technol, Sch Mech Engn, Hengam St, Tehran 1684613114, Iran
[2] Univ Tehran, Fac New Sci & Technol, North Kargar St, Tehran 1439957131, Iran
[3] Amirkabir Univ Technol, Dept Aerosp Engn, Hafez Ave, Tehran 1591634311, Iran
关键词
Visual object tracking; Object tracking dataset; Aerial robot; Deep learning; Visual-inertial navigation;
D O I
10.1016/j.eswa.2023.120483
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Visual object tracking for autonomy of aerial robots could become challenging especially in the presence of target or camera fast motions and long-term occlusions. This paper presents a visual-inertial tracking paradigm by incorporating the camera kinematics states into the visual object tracking pipelines. We gathered a dataset of image sequences with the addition of camera's position and orientation measurements as well as the object's position measurement. For the cases of long-term object occlusion, we provide ground-truth boxes derived from mapping the measured object position onto the image frame. A search zone proposal method is developed based on the estimation of object future position represented in the inertial frame and projected back onto the image frame using the camera states. This search zone, which is robust to fast camera/target motions, is fused into the original search zone settings of the base tracker. Also proposed is a measure indicating the confidence of a tracking structure in keeping track of a correct target and reporting the tracking failure in-time. Accordingly, the model updating mechanism of base tracker is modulated to avoid recovering of wrong objects as the target. The proposed modifications are benchmarked on nine visual object tracking algorithms including five state-of-art deep learning structures, namely DiMP, PrDiMP, KYS, ToMP, and MixFormer. Most of the trackers are remarkably improved by the modifications with up to 8% increase in precision. Modified PrDiMP tracker yields the best precision of 68.4%, more than all considered original (and modified) trackers. Source code and dataset are made available online.1
引用
收藏
页数:15
相关论文
共 50 条
  • [1] VIPose: Real-time Visual-Inertial 6D Object Pose Tracking
    Ge, Rundong
    Loianno, Giuseppe
    2021 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2021, : 4597 - 4603
  • [2] Object-Based Visual-Inertial Tracking for Additive Fabrication
    Sandy T.
    Buchli J.
    IEEE Robotics and Automation Letters, 2018, 3 (03) : 1370 - 1377
  • [3] Learning to Fuse: A Deep Learning Approach to Visual-Inertial Camera Pose Estimation
    Rambach, Jason R.
    Tewari, Aditya
    Pagani, Alain
    Stricker, Didier
    PROCEEDINGS OF THE 2016 15TH IEEE INTERNATIONAL SYMPOSIUM ON MIXED AND AUGMENTED REALITY (ISMAR), 2016, : 71 - 76
  • [4] BodySLAM plus plus : Fast and Tightly-Coupled Visual-Inertial Camera and Human Motion Tracking
    Henning, Dorian F.
    Choi, Christopher
    Schaefer, Simon
    Leutenegger, Stefan
    2023 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, IROS, 2023, : 3781 - 3788
  • [5] A Visual-Inertial Servoing Method for Tracking Object with Two Landmarks and an Inertial Measurement Unit
    Nguyen, Ho-Quoc-Phuong
    Kang, Hee-Jun
    Suh, Young-Soo
    INTERNATIONAL JOURNAL OF CONTROL AUTOMATION AND SYSTEMS, 2011, 9 (02) : 317 - 327
  • [6] A visual-inertial servoing method for tracking object with two landmarks and an inertial measurement unit
    Ho-Quoc-Phuong Nguyen
    Hee-Jun Kang
    Young-Soo Suh
    International Journal of Control, Automation and Systems, 2011, 9 : 317 - 327
  • [7] Schmidt-EKF-based Visual-Inertial Moving Object Tracking
    Eckenhoff, Kevin
    Geneva, Patrick
    Merrill, Nathaniel
    Huang, Guoquan
    2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 651 - 657
  • [8] A Visual-Inertial Dynamic Object Tracking SLAM Tightly Coupled System
    Zhang, Hanxuan
    Wang, Dingyi
    Huo, Ju
    IEEE SENSORS JOURNAL, 2023, 23 (17) : 19905 - 19917
  • [9] Pose estimation by Omnidirectional Visual-Inertial Odometry
    Ramezani, Milad
    Khoshelham, Kourosh
    Fraser, Clive
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2018, 105 : 26 - 37
  • [10] Visual-Inertial Object Detection and Mapping
    Fei, Xiaohan
    Soatto, Stefano
    COMPUTER VISION - ECCV 2018, PT XI, 2018, 11215 : 318 - 334