Compressing Deep Neural Networks for Recognizing Places

被引:0
|
作者
Saha, Soham [1 ]
Varma, Girish [1 ]
Jawahar, C. V. [1 ]
机构
[1] Int Inst Informat Technol, KCIS, CVIT, Hyderabad, India
关键词
Visual Place Recognition; Model Compression; Image Retrieval;
D O I
10.1109/ACPR.2017.154
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Visual place recognition on low memory devices such as mobile phones and robotics systems is a challenging problem. The state of the art models for this task uses deep learning architectures having close to 100 million parameters which takes over 400MB of memory. This makes these models infeasible to be deployed in low memory devices and gives rise to the need of compressing them. Hence we study the effectiveness of model compression techniques like trained quantization and pruning for reducing the number of parameters on one of the best performing image retrieval models called NetVLAD. We show that a compressed network can be created by starting with a model pre-trained for the task of visual place recognition and then fine-tuning it via trained pruning and quantization. The compressed model is able to produce the same mAP as the original uncompressed network. We achieve almost 50% parameter pruning with no loss in mAP and 70% pruning with close to 2% mAP reduction, while also performing 8-bit quantization. Furthermore, together with 5-bit quantization, we perform about 50% parameter reduction by pruning and get only about 3% reduction in mAP. The resulting compressed networks have sizes of around 30MB and 65MB which makes them easily usable in memory constrained devices.
引用
收藏
页码:352 / 357
页数:6
相关论文
共 50 条
  • [11] Progressive principle component analysis for compressing deep convolutional neural networks
    Zhou, Jing
    Qi, Haobo
    Chen, Yu
    Wang, Hansheng
    NEUROCOMPUTING, 2021, 440 : 197 - 206
  • [12] Accelerating and Compressing Deep Neural Networks for Massive MIMO CSI Feedback
    Erak, Omar
    Abou-Zeid, Hatem
    ICC 2023-IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS, 2023, : 1029 - 1035
  • [13] Compressing Deep Graph Neural Networks via Adversarial Knowledge Distillation
    He, Huarui
    Wang, Jie
    Zhang, Zhanqiu
    Wu, Feng
    PROCEEDINGS OF THE 28TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2022, 2022, : 534 - 544
  • [14] A Knee-Guided Evolutionary Algorithm for Compressing Deep Neural Networks
    Zhou, Yao
    Yen, Gary G.
    Yi, Zhang
    IEEE TRANSACTIONS ON CYBERNETICS, 2021, 51 (03) : 1626 - 1638
  • [15] Deep Neural Networks for Recognizing Online Handwritten Mathematical Symbols
    Nguyen, Hai Dai
    Le, Anh Duc
    Nakagawa, Masaki
    PROCEEDINGS 3RD IAPR ASIAN CONFERENCE ON PATTERN RECOGNITION ACPR 2015, 2015, : 121 - 125
  • [16] Compressing DMA Engine: Leveraging Activation Sparsity for Training Deep Neural Networks
    Rhu, Minsoo
    O'Connor, Mike
    Chatterjee, Niladrish
    Pool, Jeff
    Kwon, Youngeun
    Keckler, Stephen W.
    2018 24TH IEEE INTERNATIONAL SYMPOSIUM ON HIGH PERFORMANCE COMPUTER ARCHITECTURE (HPCA), 2018, : 78 - 91
  • [17] Bit-Quantized-Net: An Effective Method for Compressing Deep Neural Networks
    Chunshan Li
    Qing Du
    Xiaofei Xu
    Jinhui Zhu
    Dianhui Chu
    Mobile Networks and Applications, 2021, 26 : 104 - 113
  • [18] Bit-Quantized-Net: An Effective Method for Compressing Deep Neural Networks
    Li, Chunshan
    Du, Qing
    Xu, Xiaofei
    Zhu, Jinhui
    Chu, Dianhui
    MOBILE NETWORKS & APPLICATIONS, 2021, 26 (01): : 104 - 113
  • [19] A Novel Low-Bit Quantization Strategy for Compressing Deep Neural Networks
    Long, Xin
    Zeng, XiangRong
    Ben, Zongcheng
    Zhou, Dianle
    Zhang, Maojun
    COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2020, 2020 (2020)
  • [20] Compressing fully connected layers of deep neural networks using permuted features
    Nagaraju, Dara
    Chandrachoodan, Nitin
    IET COMPUTERS AND DIGITAL TECHNIQUES, 2023, 17 (3-4): : 149 - 161