Robust fusion for RGB-D tracking using CNN features

被引:17
|
作者
Wang, Yong [1 ,2 ]
Wei, Xian [3 ]
Shen, Hao [4 ,5 ]
Ding, Lu [6 ]
Wan, Jiuqing [7 ]
机构
[1] Sun Yat Sen Univ, Sch Aeronaut & Astronaut, Guangzhou, Guangdong, Peoples R China
[2] Univ Ottawa, Sch Elect & Comp Sci, Ottawa, ON, Canada
[3] Chinese Acad Sci, Fujian Inst Res Struct Matter, Fuzhou, Peoples R China
[4] Tech Univ Munich, Munich, Germany
[5] Fortiss GmbH, Munich, Germany
[6] Shanghai Jiao Tong Univ, Sch Aeronaut & Astronaut, Shanghai 200240, Peoples R China
[7] Beijing Univ Aeronaut & Astronaut, Dept Automat, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
RGB-D tracking; Robust fusion; Hierarchical convolutional neural network; Correlation filter tracking; DEEP CONVOLUTIONAL NETWORKS; VISUAL TRACKING; OBJECT TRACKING; MODEL; TIME;
D O I
10.1016/j.asoc.2020.106302
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, RGB-D sensors have become popular. Many computer vision problems can be better dealt with depth data. It is a challenging problem to integrate depth data into a visual object tracker to address the problems such as scale change and occlusion. In this paper, we propose a robust fusion based RGB-D tracking method. Specifically, hierarchical convolutional neural network (CNN) features are first adopted to encode RGB and depth images separately. Next, target is tracked based on correlation filter tracking framework. Then the results of each CNN feature are localized according to the tracking results in a short period of time. Finally, the target is localized by jointly fusing the results of RGB and depth images. Model updating is finally carried out according to the differences between RGB and depth images. Experiments on the University of Birmingham RGB-D Tracking Benchmark (BTB) and the Princeton RGB-D Tracking Benchmark (PTB) achieve comparable results to state-of-the-art methods. (C) 2020 Elsevier B.V. All rights reserved.
引用
收藏
页数:9
相关论文
共 50 条
  • [41] Robust approach to inverse lighting using RGB-D images
    Choe, Junsuk
    Shim, Hyunjung
    INFORMATION SCIENCES, 2018, 438 : 73 - 94
  • [42] Robust 6D Object Pose Estimation by Learning RGB-D Features
    Tian, Meng
    Pan, Liang
    Ang, Marcelo H., Jr.
    Lee, Gim Hee
    2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 6218 - 6224
  • [43] A Fusion Network for Semantic Segmentation Using RGB-D Data
    Yuan, Jiahui
    Zhang, Kun
    Xia, Yifan
    Qi, Lin
    Dong, Junyu
    NINTH INTERNATIONAL CONFERENCE ON GRAPHIC AND IMAGE PROCESSING (ICGIP 2017), 2018, 10615
  • [44] Tracking Spatially Distributed Features in KLT Algorithms for RGB-D Visual Odometry
    da Silva, Bruno Marques F.
    Correia, Luiz Felipe M.
    Bezerra, Kallil de A.
    Goncalves, Luiz Marcos G.
    2017 WORKSHOP OF COMPUTER VISION (WVC), 2017, : 67 - 72
  • [45] Exploiting Multi-layer Features Using a CNN-RNN Approach for RGB-D Object Recognition
    Caglayan, Ali
    Can, Ahmet Burak
    COMPUTER VISION - ECCV 2018 WORKSHOPS, PT III, 2019, 11131 : 675 - 688
  • [46] Accurate and Robust RGB-D Visual Odometry Based on Point and Line Features
    Zhao, Guojie
    Zhang, Yupeng
    Liu, Peichu
    Wu, Haoen
    Cui, Mingyang
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, KSEM 2021, PT II, 2021, 12816 : 500 - 510
  • [47] Human Action Recognition Using RGB-D Image Features
    Tang C.
    Wang W.
    Zhang C.
    Peng H.
    Li W.
    Moshi Shibie yu Rengong Zhineng/Pattern Recognition and Artificial Intelligence, 2019, 32 (10): : 901 - 908
  • [48] 3D Localization of Hand Acupoints Using Hand Geometry and Landmark Points Based on RGB-D CNN Fusion
    Danish Masood
    Jiang Qi
    Annals of Biomedical Engineering, 2022, 50 : 1103 - 1115
  • [49] 3D Localization of Hand Acupoints Using Hand Geometry and Landmark Points Based on RGB-D CNN Fusion
    Masood, Danish
    Qi, Jiang
    ANNALS OF BIOMEDICAL ENGINEERING, 2022, 50 (09) : 1103 - 1115
  • [50] Depth-Aware CNN for RGB-D Segmentation
    Wang, Weiyue
    Neumann, Ulrich
    COMPUTER VISION - ECCV 2018, PT XI, 2018, 11215 : 144 - 161