Learning image representations tied to ego-motion

被引:109
|
作者
Jayaraman, Dinesh [1 ]
Grauman, Kristen [1 ]
机构
[1] Univ Texas Austin, Austin, TX 78712 USA
关键词
D O I
10.1109/ICCV.2015.166
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Understanding how images of objects and scenes behave in response to specific ego-motions is a crucial aspect of proper visual development, yet existing visual learning methods are conspicuously disconnected from the physical source of their images. We propose to exploit proprioceptive motor signals to provide unsupervised regularization in convolutional neural networks to learn visual representations from egocentric video. Specifically, we enforce that our learned features exhibit equivariance i.e. they respond predictably to transformations associated with distinct ego-motions. With three datasets, we show that our unsupervised feature learning approach significantly outperforms previous approaches on visual recognition and next-best-view prediction tasks. In the most challenging test, we show that features learned from video captured on an autonomous driving platform improve large-scale scene recognition in static images from a disjoint domain.
引用
收藏
页码:1413 / 1421
页数:9
相关论文
共 50 条
  • [1] Unsupervised Visual Ego-motion Learning for Robots
    Khalilbayli, Fidan
    Bayram, Baris
    Ince, Gokhan
    2019 4TH INTERNATIONAL CONFERENCE ON COMPUTER SCIENCE AND ENGINEERING (UBMK), 2019, : 676 - 681
  • [2] Towards Visual Ego-motion Learning in Robots
    Pillai, Sudeep
    Leonard, John J.
    2017 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2017, : 5533 - 5540
  • [3] Adversarial Learning for Joint Optimization of Depth and Ego-Motion
    Wang, Anjie
    Fang, Zhijun
    Gao, Yongbin
    Tan, Songchao
    Wang, Shanshe
    Ma, Siwei
    Hwang, Jenq-Neng
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 : 4130 - 4142
  • [4] Unsupervised Learning of Depth and Ego-Motion from Video
    Zhou, Tinghui
    Brown, Matthew
    Snavely, Noah
    Lowe, David G.
    30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 6612 - +
  • [5] Sparse Representations for Object- and Ego-Motion Estimations in Dynamic Scenes
    Kashyap, Hirak J.
    Fowlkes, Charless C.
    Krichmar, Jeffrey L.
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (06) : 2521 - 2534
  • [6] Estimating Ego-Motion in Panoramic Image Sequences with Inertial Measurements
    Schill, Felix
    Mahony, Robert
    Corke, Peter
    ROBOTICS RESEARCH, 2011, 70 : 87 - +
  • [7] Ego-motion and omnidirectional cameras
    Gluckman, J
    Nayar, SK
    SIXTH INTERNATIONAL CONFERENCE ON COMPUTER VISION, 1998, : 999 - 1005
  • [8] PoseConvGRU: A Monocular Approach for Visual Ego-motion Estimation by Learning
    Zhai, Guangyao
    Liu, Liang
    Zhang, Linjian
    Liu, Yong
    Jiang, Yunliang
    PATTERN RECOGNITION, 2020, 102
  • [9] Unsupervised monocular depth and ego-motion learning with structure and semantics
    Casser, Vincent
    Pirk, Soeren
    Mahjourian, Reza
    Angelova, Anelia
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2019), 2019, : 381 - 388
  • [10] An Ego-Motion Detection System Employing Directional-Edge-Based Motion Field Representations
    Hao, Jia
    Shibata, Tadashi
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2010, E93D (01): : 94 - 106