Compressing Deep Neural Networks for Recognizing Places

被引:0
|
作者
Saha, Soham [1 ]
Varma, Girish [1 ]
Jawahar, C. V. [1 ]
机构
[1] Int Inst Informat Technol, KCIS, CVIT, Hyderabad, India
关键词
Visual Place Recognition; Model Compression; Image Retrieval;
D O I
10.1109/ACPR.2017.154
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Visual place recognition on low memory devices such as mobile phones and robotics systems is a challenging problem. The state of the art models for this task uses deep learning architectures having close to 100 million parameters which takes over 400MB of memory. This makes these models infeasible to be deployed in low memory devices and gives rise to the need of compressing them. Hence we study the effectiveness of model compression techniques like trained quantization and pruning for reducing the number of parameters on one of the best performing image retrieval models called NetVLAD. We show that a compressed network can be created by starting with a model pre-trained for the task of visual place recognition and then fine-tuning it via trained pruning and quantization. The compressed model is able to produce the same mAP as the original uncompressed network. We achieve almost 50% parameter pruning with no loss in mAP and 70% pruning with close to 2% mAP reduction, while also performing 8-bit quantization. Furthermore, together with 5-bit quantization, we perform about 50% parameter reduction by pruning and get only about 3% reduction in mAP. The resulting compressed networks have sizes of around 30MB and 65MB which makes them easily usable in memory constrained devices.
引用
收藏
页码:352 / 357
页数:6
相关论文
共 50 条
  • [1] Anonymous Model Pruning for Compressing Deep Neural Networks
    Zhang, Lechun
    Chen, Guangyao
    Shi, Yemin
    Zhang, Quan
    Tan, Mingkui
    Wang, Yaowei
    Tian, Yonghong
    Huang, Tiejun
    THIRD INTERNATIONAL CONFERENCE ON MULTIMEDIA INFORMATION PROCESSING AND RETRIEVAL (MIPR 2020), 2020, : 161 - 164
  • [2] COMPRESSING DEEP NEURAL NETWORKS FOR EFFICIENT SPEECH ENHANCEMENT
    Tan, Ke
    Wang, DeLiang
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 8358 - 8362
  • [3] CUP: Cluster Pruning for Compressing Deep Neural Networks
    Duggal, Rahul
    Xiao, Cao
    Vuduc, Richard
    Duen Horng Chau
    Sun, Jimeng
    2021 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2021, : 5102 - 5106
  • [4] Compressing Deep Neural Networks With Sparse Matrix Factorization
    Wu, Kailun
    Guo, Yiwen
    Zhang, Changshui
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (10) : 3828 - 3838
  • [5] COMPRESSING DEEP NEURAL NETWORKS FOR EFFICIENT VISUAL INFERENCE
    Ge, Shiming
    Luo, Zhao
    Zhao, Shengwei
    Jin, Xin
    Zhang, Xiao-Yu
    2017 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2017, : 667 - 672
  • [6] Compressing deep neural networks by matrix product operators
    Gao, Ze-Feng
    Cheng, Song
    He, Rong-Qiang
    Xie, Z. Y.
    Zhao, Hui-Hai
    Lu, Zhong-Yi
    Xiang, Tao
    PHYSICAL REVIEW RESEARCH, 2020, 2 (02):
  • [7] A Unified Approximation Framework for Compressing and Accelerating Deep Neural Networks
    Ma, Yuzhe
    Chen, Ran
    Li, Wei
    Shang, Fanhua
    Yu, Wenjian
    Cho, Minsik
    Yu, Bei
    2019 IEEE 31ST INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2019), 2019, : 376 - 383
  • [8] Compressing deep-quaternion neural networks with targeted regularisation
    Vecchi, Riccardo
    Scardapane, Simone
    Comminiello, Danilo
    Uncini, Aurelio
    CAAI TRANSACTIONS ON INTELLIGENCE TECHNOLOGY, 2020, 5 (03) : 172 - 176
  • [9] Small Is Beautiful: Compressing Deep Neural Networks for Partial Domain Adaptation
    Ma, Yuzhe
    Yao, Xufeng
    Chen, Ran
    Li, Ruiyu
    Shen, Xiaoyong
    Yu, Bei
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (03) : 3575 - 3585
  • [10] Compressing Deep Neural Networks using a Rank-Constrained Topology
    Nakkiran, Preetum
    Alvarez, Raziel
    Prabhavalkar, Rohit
    Parada, Carolina
    16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5, 2015, : 1473 - 1477