Visual-inertial object tracking: Incorporating camera pose into motion models

被引:3
|
作者
Shahbazi, Mohammad [1 ]
Mirtajadini, Seyed Hojat [2 ]
Fahimi, Hamidreza [3 ]
机构
[1] Iran Univ Sci & Technol, Sch Mech Engn, Hengam St, Tehran 1684613114, Iran
[2] Univ Tehran, Fac New Sci & Technol, North Kargar St, Tehran 1439957131, Iran
[3] Amirkabir Univ Technol, Dept Aerosp Engn, Hafez Ave, Tehran 1591634311, Iran
关键词
Visual object tracking; Object tracking dataset; Aerial robot; Deep learning; Visual-inertial navigation;
D O I
10.1016/j.eswa.2023.120483
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Visual object tracking for autonomy of aerial robots could become challenging especially in the presence of target or camera fast motions and long-term occlusions. This paper presents a visual-inertial tracking paradigm by incorporating the camera kinematics states into the visual object tracking pipelines. We gathered a dataset of image sequences with the addition of camera's position and orientation measurements as well as the object's position measurement. For the cases of long-term object occlusion, we provide ground-truth boxes derived from mapping the measured object position onto the image frame. A search zone proposal method is developed based on the estimation of object future position represented in the inertial frame and projected back onto the image frame using the camera states. This search zone, which is robust to fast camera/target motions, is fused into the original search zone settings of the base tracker. Also proposed is a measure indicating the confidence of a tracking structure in keeping track of a correct target and reporting the tracking failure in-time. Accordingly, the model updating mechanism of base tracker is modulated to avoid recovering of wrong objects as the target. The proposed modifications are benchmarked on nine visual object tracking algorithms including five state-of-art deep learning structures, namely DiMP, PrDiMP, KYS, ToMP, and MixFormer. Most of the trackers are remarkably improved by the modifications with up to 8% increase in precision. Modified PrDiMP tracker yields the best precision of 68.4%, more than all considered original (and modified) trackers. Source code and dataset are made available online.1
引用
收藏
页数:15
相关论文
共 50 条
  • [41] A Visual-inertial Fusion Based Tracking System for Mobile Augmented Reality
    Lin, Cheng
    Wang, Lianghao
    Li, Dongxiao
    Zhang, Ming
    2015 IEEE INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING, COMMUNICATIONS AND COMPUTING (ICSPCC), 2015, : 956 - 960
  • [42] Real-Time Motion Tracking for Mobile Augmented/Virtual Reality Using Adaptive Visual-Inertial Fusion
    Fang, Wei
    Zheng, Lianyu
    Deng, Huanjun
    Zhang, Hongbo
    SENSORS, 2017, 17 (05)
  • [43] 360-VIO: A Robust Visual-Inertial Odometry Using a 360° Camera
    Wu, Qi
    Xu, Xiangyu
    Chen, Xieyuanli
    Pei, Ling
    Long, Chao
    Deng, Junyuan
    Liu, Guoqing
    Yang, Sheng
    Wen, Shilei
    Yu, Wenxian
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2024, 71 (09) : 11136 - 11145
  • [44] Monocular Visual-Inertial Fusion with Online Initialization and Camera-IMU Calibration
    Yang, Zhenfei
    Shen, Shaojie
    2015 IEEE INTERNATIONAL SYMPOSIUM ON SAFETY, SECURITY, AND RESCUE ROBOTICS (SSRR), 2015,
  • [45] Closed-Form Solution of Visual-Inertial Structure from Motion
    Martinelli, Agostino
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2014, 106 (02) : 138 - 152
  • [46] Motion based object tracking with mobile camera
    Electronics and Telecommunications, Research Inst, Daejon, Korea, Republic of
    Electron Lett, 3 (256-258):
  • [47] Motion based object tracking with mobile camera
    Lee, KW
    Ryu, SW
    Lee, SJ
    Park, KT
    ELECTRONICS LETTERS, 1998, 34 (03) : 256 - 258
  • [48] Multi-Camera Visual-Inertial Navigation with Online Intrinsic and Extrinsic Calibration
    Eckenhoff, Kevin
    Geneva, Patrick
    Bloecker, Jesse
    Huang, Guoquan
    2019 INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2019, : 3158 - 3164
  • [49] Visual-inertial navigation algorithm development using photorealistic camera simulation in the loop
    Sayre-McCord, Thomas
    Guerra, Winter
    Antonini, Amado
    Arneberg, Jasper
    Brown, Austin
    Cavalheiro, Guilherme
    Fang, Yajun
    Gorodetsky, Alex
    McCoy, Dave
    Quilter, Sebastian
    Riether, Fabian
    Tal, Ezra
    Terzioglu, Yunus
    Carlone, Luca
    Karaman, Sertac
    2018 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2018, : 2566 - 2573
  • [50] Efficient Multi-Camera Visual-Inertial SLAM for Micro Aerial Vehicles
    Houben, Sebastian
    Quenzel, Jan
    Krombach, Nicola
    Behnke, Sven
    2016 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS 2016), 2016, : 1616 - 1622