Learning Deep Features for Scene Recognition using Places Database

被引:0
|
作者
Zhou, Bolei [1 ]
Lapedriza, Agata [1 ,3 ]
Xiao, Jianxiong [2 ]
Totralba, Antonio [1 ]
Oliva, Aude [1 ]
机构
[1] MIT, Cambridge, MA 02139 USA
[2] Princeton Univ, Princeton, NJ 08544 USA
[3] Univ Oberta Catalunya, Barcelona, Spain
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Scene recognition is one of the hallmark tasks of computer vision, allowing definition of a context for object recognition. Whereas the tremendous recent progress in object recognition tasks is due to the availability of large datasets like ImageNet and the rise of Convolutional Neural Networks (CNNs) for learning high-level features, performance at scene recognition has not attained the same level of success. This may be because current deep features trained from ImageNet are not competitive enough for such tasks. Here, we introduce a new scene-centric database called Places with over 7 million labeled pictures of scenes. We propose new methods to compare the density and diversity of image datasets and show that Places is as dense as other scene datasets and has more diversity. Using CNN, we learn deep features for scene recognition tasks, and establish new state-of-the-art results on several scene-centric datasets. A visualization of the CNN layers' responses allows us to show differences in the internal representations of object-centric and scene-centric networks.
引用
收藏
页数:9
相关论文
共 50 条
  • [21] Deep Learning Based Application for Indoor Scene Recognition
    Afif, Mouna
    Ayachi, Riadh
    Said, Yahia
    Atri, Mohamed
    NEURAL PROCESSING LETTERS, 2020, 51 (03) : 2827 - 2837
  • [22] Dictionary Learning Inspired Deep Network for Scene Recognition
    Liu, Yang
    Chen, Qingchao
    Chen, Wei
    Wassell, Ian
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 7178 - 7185
  • [23] Deep transformation learning for face recognition in the unconstrained scene
    Chen, Guanhao
    Shao, Yanqing
    Tang, Chaowei
    Jin, Zhuoyi
    Zhang, Jinkun
    MACHINE VISION AND APPLICATIONS, 2018, 29 (03) : 513 - 523
  • [24] Fusing Object Semantics and Deep Appearance Features for Scene Recognition
    Sun, Ning
    Li, Wenli
    Liu, Jixin
    Han, Guang
    Wu, Cong
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2019, 29 (06) : 1715 - 1728
  • [25] Scene Text Detection and Recognition: The Deep Learning Era
    Shangbang Long
    Xin He
    Cong Yao
    International Journal of Computer Vision, 2021, 129 : 161 - 184
  • [26] Scene Text Detection and Recognition: The Deep Learning Era
    Long, Shangbang
    He, Xin
    Yao, Cong
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2021, 129 (01) : 161 - 184
  • [27] Scene Recognition with Convolutional Residual Features via Deep Forest
    Han, Mingfei
    Li, Shengyang
    Wan, Xue
    Liu, Guiyang
    2018 IEEE 3RD INTERNATIONAL CONFERENCE ON IMAGE, VISION AND COMPUTING (ICIVC), 2018, : 178 - 182
  • [28] Texture Gradient and Deep Features Fusion-Based Image Scene Geometry Recognition System Using Extreme Learning Machine
    Khan, Altaf
    Chefranov, Alexander G.
    Demirel, Hasan
    2020 THE 3RD INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTICS AND CONTROL ENGINEERING (IRCE 2020), 2020, : 37 - 41
  • [29] Pattern recognition and features selection for speech emotion recognition model using deep learning
    Kittisak Jermsittiparsert
    Abdurrahman Abdurrahman
    Parinya Siriattakul
    Ludmila A. Sundeeva
    Wahidah Hashim
    Robbi Rahim
    Andino Maseleno
    International Journal of Speech Technology, 2020, 23 : 799 - 806
  • [30] Pattern recognition and features selection for speech emotion recognition model using deep learning
    Jermsittiparsert, Kittisak
    Abdurrahman, Abdurrahman
    Siriattakul, Parinya
    Sundeeva, Ludmila A.
    Hashim, Wahidah
    Rahim, Robbi
    Maseleno, Andino
    INTERNATIONAL JOURNAL OF SPEECH TECHNOLOGY, 2020, 23 (04) : 799 - 806