Robust fusion for RGB-D tracking using CNN features

被引:17
|
作者
Wang, Yong [1 ,2 ]
Wei, Xian [3 ]
Shen, Hao [4 ,5 ]
Ding, Lu [6 ]
Wan, Jiuqing [7 ]
机构
[1] Sun Yat Sen Univ, Sch Aeronaut & Astronaut, Guangzhou, Guangdong, Peoples R China
[2] Univ Ottawa, Sch Elect & Comp Sci, Ottawa, ON, Canada
[3] Chinese Acad Sci, Fujian Inst Res Struct Matter, Fuzhou, Peoples R China
[4] Tech Univ Munich, Munich, Germany
[5] Fortiss GmbH, Munich, Germany
[6] Shanghai Jiao Tong Univ, Sch Aeronaut & Astronaut, Shanghai 200240, Peoples R China
[7] Beijing Univ Aeronaut & Astronaut, Dept Automat, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
RGB-D tracking; Robust fusion; Hierarchical convolutional neural network; Correlation filter tracking; DEEP CONVOLUTIONAL NETWORKS; VISUAL TRACKING; OBJECT TRACKING; MODEL; TIME;
D O I
10.1016/j.asoc.2020.106302
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, RGB-D sensors have become popular. Many computer vision problems can be better dealt with depth data. It is a challenging problem to integrate depth data into a visual object tracker to address the problems such as scale change and occlusion. In this paper, we propose a robust fusion based RGB-D tracking method. Specifically, hierarchical convolutional neural network (CNN) features are first adopted to encode RGB and depth images separately. Next, target is tracked based on correlation filter tracking framework. Then the results of each CNN feature are localized according to the tracking results in a short period of time. Finally, the target is localized by jointly fusing the results of RGB and depth images. Model updating is finally carried out according to the differences between RGB and depth images. Experiments on the University of Birmingham RGB-D Tracking Benchmark (BTB) and the Princeton RGB-D Tracking Benchmark (PTB) achieve comparable results to state-of-the-art methods. (C) 2020 Elsevier B.V. All rights reserved.
引用
收藏
页数:9
相关论文
共 50 条
  • [21] Multimodal Deep Feature Fusion (MMDFF) for RGB-D Tracking
    Jiang, Ming-xin
    Deng, Chao
    Zhang, Ming-min
    Shan, Jing-song
    Zhang, Haiyan
    COMPLEXITY, 2018,
  • [22] Comparison of Clustering Methods for Tracking Features in RGB-D Images
    Pancham, Ardhisha
    Withey, Daniel
    Bright, Glen
    PROCEEDINGS OF THE IECON 2016 - 42ND ANNUAL CONFERENCE OF THE IEEE INDUSTRIAL ELECTRONICS SOCIETY, 2016, : 871 - 876
  • [23] Robust RGB-D Visual Odometry Based on Planar Features
    Chen, Baifan
    Liu, Chunfa
    Tong, Yu
    Wu, Qian
    2017 IEEE 7TH ANNUAL INTERNATIONAL CONFERENCE ON CYBER TECHNOLOGY IN AUTOMATION, CONTROL, AND INTELLIGENT SYSTEMS (CYBER), 2017, : 449 - 453
  • [24] Robust RGB-D Camera Tracking using Optimal Key-frame Selection
    Han, Kyung Min
    Kim, Young J.
    2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 6275 - 6281
  • [25] Detecting and Tracking People using an RGB-D Camera via Multiple Detector Fusion
    Choi, Wongun
    Pantofaru, Caroline
    Savarese, Silvio
    2011 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCV WORKSHOPS), 2011,
  • [26] A Hybrid Real-Time Visual Tracking Using Compressive RGB-D Features
    Zhao, Mengyuan
    Luo, Heng
    Tafti, Ahmad P.
    Lin, Yuanchang
    He, Guotian
    ADVANCES IN VISUAL COMPUTING, PT I (ISVC 2015), 2015, 9474 : 561 - 573
  • [27] Robust RGB-D SLAM Using Point and Line Features for Low Textured Scene
    Zou, Yajing
    Eldemiry, Amr
    Li, Yaxin
    Chen, Wu
    SENSORS, 2020, 20 (17) : 1 - 20
  • [28] Tracking an RGB-D Camera Using Points and Planes
    Ataer-Cansizoglu, Esra
    Taguchi, Yuichi
    Ramalingam, Srikumar
    Garaas, Tyler
    2013 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW), 2013, : 51 - 58
  • [29] Robust Texture Mapping Using RGB-D Cameras
    Oliveira, Miguel
    Lim, Gi-Hyun
    Madeira, Tiago
    Dias, Paulo
    Santos, Vitor
    SENSORS, 2021, 21 (09)
  • [30] RGB-D Indoor Mapping Using Deep Features
    Guclu, Oguzhan
    Caglayan, Adi
    Can, Ahmet Burak
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2019), 2019, : 1248 - 1257