Visual Co-occurrence Network: Using Context for Large-Scale Object Recognition in Retail

被引:0
|
作者
Advani, Siddharth [1 ]
Smith, Brigid [1 ]
Tanabe, Yasuki [2 ]
Irick, Kevin [3 ]
Cotter, Matthew [1 ]
Sampson, Jack [1 ]
Narayanan, Vijaykrishnan [1 ]
机构
[1] Penn State Univ, University Pk, PA 16802 USA
[2] Toshiba, Tokyo, Japan
[3] SiliconScapes LLC, State Coll, PA USA
关键词
D O I
暂无
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
In any visual object recognition system, the classification accuracy will likely determine the usefulness of the system as a whole. In many real-world applications, it is also important to be able to recognize a large number of diverse objects for the system to be robust enough to handle the sort of tasks that the human visual system handles on an average day. These objectives are often at odds with performance, as running too large of a number of detectors on any one scene will be prohibitively slow for use in any real-time scenario. However, visual information has temporal and spatial context that can be exploited to reduce the number of detectors that need to be triggered at any given instance. In this paper, we propose a dynamic approach to encode such context, called Visual Co-occurrence Network (ViCoNet) that establishes relationships between objects observed in a visual scene. We investigate the utility of ViCoNet when integrated into a vision pipeline targeted for retail shopping. When evaluated on a large and deep dataset, we achieve a 50% improvement in performance and a 7% improvement in accuracy in the best case, and a 45% improvement in performance and a 3% improvement in accuracy in the average case over an established baseline. The memory overhead of ViCoNet is around 10KB, highlighting its effectiveness on temporal big data.
引用
收藏
页码:103 / 112
页数:10
相关论文
共 50 条
  • [31] Robust object recognition using a color co-occurrence histogram and the spatial relations of image patches
    Bang, Heebeom
    Lee, Sanghoon
    Yu, Dongjin
    Suh, Il Hong
    ARTIFICIAL LIFE AND ROBOTICS, 2009, 13 (02) : 488 - 492
  • [32] SUPERPIXEL CONTEXT DESCRIPTION BASED ON VISUAL WORDS CO-OCCURRENCE MATRIX
    Santana, Tiago M. H. C.
    Torres, Ricardo da S.
    dos Santos, Jefersson A.
    IGARSS 2018 - 2018 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM, 2018, : 6199 - 6202
  • [33] Text filtering for harmful document classification method using three words co-occurrence and large-scale data processing
    Otsuka, Takanobu
    Deng, Deyue
    Ito, Takayuki
    Otsuka, T. (otsuka.takanobu@nitech.ac.jp), 1600, Institute of Electrical Engineers of Japan (134): : 168 - 175
  • [34] Automatic visual inspection using the co-occurrence approach
    Tobias, OJ
    Seara, R
    Soares, FAP
    Bermudez, JCM
    38TH MIDWEST SYMPOSIUM ON CIRCUITS AND SYSTEMS, PROCEEDINGS, VOLS 1 AND 2, 1996, : 154 - 157
  • [35] Text Filtering for Harmful Document Classification Using Three-Word Co-Occurrence and Large-Scale Data Processing
    Otsuka, Takanobu
    Deng, Deyue
    Ito, Takayuki
    ELECTRONICS AND COMMUNICATIONS IN JAPAN, 2015, 98 (10) : 31 - 40
  • [36] A Convolutional Neural Network to Perform Object Detection and Identification in Visual Large-Scale Data
    Ayachi, Riadh
    Said, Yahia
    Atri, Mohamed
    BIG DATA, 2021, 9 (01) : 41 - 52
  • [37] Fuzzy relational distance for large-scale object recognition
    Huet, B
    Hancock, ER
    1998 IEEE COMPUTER SOCIETY CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, PROCEEDINGS, 1998, : 138 - 143
  • [38] Visual analysis of large-scale network anomalies
    Liao, Q.
    Shi, L.
    Wang, C.
    IBM JOURNAL OF RESEARCH AND DEVELOPMENT, 2013, 57 (3-4)
  • [39] Co-occurrence patterns and the large-scale spatial structure of benthic communities in seagrass meadows and bare sand
    Kraan, Casper
    Thrush, Simon F.
    Dormann, Carsten F.
    BMC ECOLOGY, 2020, 20 (01)
  • [40] Co-occurrence Features and Neural Network Classification Approach for Iris Recognition
    Vyas, Kitesh
    Kanumuri, Tirupathiraju
    Sheoran, Gyanendra
    Duffey, Pawan
    2017 FOURTH INTERNATIONAL CONFERENCE ON IMAGE INFORMATION PROCESSING (ICIIP), 2017, : 306 - 311