Compressing Deep Model With Pruning and Tucker Decomposition for Smart Embedded Systems

被引:18
|
作者
Dai, Cheng [1 ]
Liu, Xingang [2 ]
Cheng, Hongqiang [2 ]
Yang, Laurence T. [3 ]
Deen, M. Jamal [4 ]
机构
[1] Sichuan Univ, Coll Comp Sci, Chengdu 610017, Peoples R China
[2] Univ Elect Sci & Technol China, Sch Informat & Commun Engn, Chengdu 611731, Peoples R China
[3] St Francis Xavier Univ, Dept Comp Sci, Antigonish, NS B2G 2W5, Canada
[4] McMaster Univ, Dept Elect Engn & Comp Sci, Hamilton, ON L8S 4K1, Canada
基金
中国国家自然科学基金;
关键词
Computational modeling; Deep learning; Bayes methods; Internet of Things; Data models; Streaming media; Edge computing; Deep model compression; parameter pruning; smart embedded systems; Tucker decomposition (TD); IOT;
D O I
10.1109/JIOT.2021.3116316
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning has been proved to be one of the most effective method in feature encoding for different intelligent applications such as video-based human action recognition. However, its nonconvex optimization mechanism leads large memory consumption, which hinders its deployment on the smart embedded systems with limited computational resources. To overcome this challenge, we propose a novel deep model compression technique for smart embedded systems, which realizes both the memory size reduction and inference complexity decrease within a small drop of accuracy. First, we propose an improved naive Bayes inference-based channel parameter pruning to obtain a sparse model with higher accuracy. Then, to improve the inference efficiency, the improved Tucker decomposition method is proposed, where an improved genetic algorithm is used to optimize the Tucker ranks. Finally, to elevate the effectiveness of our proposed method, extensive experiments are conducted. The experimental results show that our method can achieve the state-of-the-art performance compared with existing methods in terms of accuracy, parameter compression, and floating-point operations reduction.
引用
收藏
页码:14490 / 14500
页数:11
相关论文
共 50 条
  • [1] ADA-Tucker: Compressing deep neural networks via adaptive dimension adjustment tucker decomposition
    Zhong, Zhisheng
    Wei, Fangyin
    Lin, Zhouchen
    Zhang, Chao
    NEURAL NETWORKS, 2019, 110 : 104 - 115
  • [2] Anonymous Model Pruning for Compressing Deep Neural Networks
    Zhang, Lechun
    Chen, Guangyao
    Shi, Yemin
    Zhang, Quan
    Tan, Mingkui
    Wang, Yaowei
    Tian, Yonghong
    Huang, Tiejun
    THIRD INTERNATIONAL CONFERENCE ON MULTIMEDIA INFORMATION PROCESSING AND RETRIEVAL (MIPR 2020), 2020, : 161 - 164
  • [3] CUP: Cluster Pruning for Compressing Deep Neural Networks
    Duggal, Rahul
    Xiao, Cao
    Vuduc, Richard
    Duen Horng Chau
    Sun, Jimeng
    2021 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2021, : 5102 - 5106
  • [4] Compressing Code for Embedded Systems
    Dias, Wanderson Roger Azevedo
    Moreno, Edward David
    2014 21ST IEEE INTERNATIONAL CONFERENCE ON ELECTRONICS, CIRCUITS AND SYSTEMS (ICECS), 2014, : 271 - 274
  • [5] Compressing convolutional neural networks with hierarchical Tucker-2 decomposition
    Gabor, Mateusz
    Zdunek, Rafal
    APPLIED SOFT COMPUTING, 2023, 132
  • [6] Big Data Compression of Smart Distribution Systems Based on Tensor Tucker Decomposition
    Zhao H.
    Ma L.
    Zhongguo Dianji Gongcheng Xuebao/Proceedings of the Chinese Society of Electrical Engineering, 2019, 39 (16): : 4744 - 4752
  • [7] Transfer channel pruning for compressing deep domain adaptation models
    Yu, Chaohui
    Wang, Jindong
    Chen, Yiqiang
    Qin, Xin
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2019, 10 (11) : 3129 - 3144
  • [8] Big data cleaning model of smart grid based on Tensor Tucker decomposition
    Yin, Jun
    Zhang, Jianye
    Li, Degao
    Wang, Tianjun
    Jing, Kang
    2020 INTERNATIONAL CONFERENCE ON BIG DATA & ARTIFICIAL INTELLIGENCE & SOFTWARE ENGINEERING (ICBASE 2020), 2020, : 166 - 169
  • [9] Transfer channel pruning for compressing deep domain adaptation models
    Chaohui Yu
    Jindong Wang
    Yiqiang Chen
    Xin Qin
    International Journal of Machine Learning and Cybernetics, 2019, 10 : 3129 - 3144
  • [10] Neuron Pruning for Compressing Deep Networks Using Maxout Architectures
    Rueda, Fernando Moya
    Grzeszick, Rene
    Fink, Gernot A.
    PATTERN RECOGNITION (GCPR 2017), 2017, 10496 : 177 - 188