Adaptive Visual-Depth Fusion Transfer

被引:1
|
作者
Cai, Ziyun [1 ]
Long, Yang [2 ]
Jing, Xiao-Yuan [1 ]
Shao, Ling [3 ]
机构
[1] Nanjing Univ Posts & Telecommun, Coll Automat, Nanjing, Jiangsu, Peoples R China
[2] Univ Newcastle, Sch Comp, Open Lab, Newcastle Upon Tyne NE4 5TG, Tyne & Wear, England
[3] Inception Inst Artificial Intelligence, Abu Dhabi, U Arab Emirates
来源
关键词
RGB-D data; Domain adaptation; Visual categorization; DOMAIN; KERNEL;
D O I
10.1007/978-3-030-20870-7_4
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
While RGB-D classification task has been actively researched in recent years, most existing methods focus on the RGB-D source to target transfer task. The application of such methods cannot address the real-world scenario where the paired depth images are not hold. This paper focuses on a more flexible task that recognizes RGB test images by transferring them into the depth domain. Such a scenario retains high performance due to gaining auxiliary information but reduces the cost of pairing RGB with depth sensors at test time. Existing methods suffer from two challenges: the utilization of the additional depth features, and the domain shifting problem due to the different mechanisms between conventional RGB cameras and depth sensors. As a step towards bridging the gap, we propose a novel method called adaptive Visual-Depth Fusion Transfer (aVDFT) which can take advantage of the depth information and handle the domain distribution mismatch simultaneously. Our key novelties are: (1) a global visual-depth metric construction algorithm that can effectively align RGB and depth data structure; (2) adaptive transformed component extraction for target domain that conditioned on invariant transfer on location, scale and depth measurement. To demonstrate the effectiveness of aVDFT, we conduct comprehensive experiments on six pairs of RGB-D datasets for object recognition, scene classification and gender recognition and demonstrate state-of-the-art performance.
引用
收藏
页码:56 / 73
页数:18
相关论文
共 50 条
  • [1] Adaptive RGB Image Recognition by Visual-Depth Embedding
    Cai, Ziyun
    Long, Yang
    Shao, Ling
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2018, 27 (05) : 2471 - 2483
  • [2] LEARNING DEPTH FROM A SINGLE IMAGE USING VISUAL-DEPTH WORDS
    Kim, Sunok
    Choi, Sunghwan
    Sohn, Kwanghoon
    2015 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2015, : 1895 - 1899
  • [3] Illusory visual-depth reversal can modulate sensations of contact surface
    Yuka Igarashi
    Keiko Omori
    Tetsuya Arai
    Yasunori Aizawa
    Experimental Brain Research, 2017, 235 : 3013 - 3022
  • [4] Illusory visual-depth reversal can modulate sensations of contact surface
    Igarashi, Yuka
    Omori, Keiko
    Arai, Tetsuya
    Aizawa, Yasunori
    EXPERIMENTAL BRAIN RESEARCH, 2017, 235 (10) : 3013 - 3022
  • [5] Visual-Depth Matching Network: Deep RGB-D Domain Adaptation With Unequal Categories
    Cai, Ziyun
    Jing, Xiao-Yuan
    Shao, Ling
    IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (06) : 4623 - 4635
  • [6] Application of a Nonlinear Recursive Visual-Depth Observer Using UH60 Flight Data
    Keshavan, Jishnu
    Escobar-Alvarez, Hector
    Dimble, Kedar D.
    Humbert, J. S.
    Goerzen, Chad L.
    Whalley, Matthew S.
    JOURNAL OF GUIDANCE CONTROL AND DYNAMICS, 2016, 39 (07) : 1501 - 1512
  • [7] 基于图像关键帧的Visual-Depth Map建立方法
    马琳
    杨浩
    谭学治
    冯冠元
    哈尔滨工业大学学报, 2018, 50 (11) : 23 - 31
  • [8] A ROS Multi-Tier UAV Localization Module Based on GNSS, Inertial and Visual-Depth Data
    Antonopoulos, Angelos
    Lagoudakis, Michail G.
    Partsinevelos, Panagiotis
    DRONES, 2022, 6 (06)
  • [9] An Adaptive Fusion Algorithm for Depth Completion
    Chen, Long
    Li, Qing
    SENSORS, 2022, 22 (12)
  • [10] HIERARCHICAL DEPTH PROCESSING WITH ADAPTIVE SEARCH RANGE AND FUSION
    Lee, Zucheul
    Nguyen, Truong Q.
    2014 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2014,