Group Pruning with Group Sparse Regularization for Deep Neural Network Compression

被引:0
|
作者
Wu, Chenglu [1 ]
Pang, Wei [1 ]
Liu, Hao [1 ]
Lu, Shengli [1 ]
机构
[1] Southeast Univ, Natl ASIC Syst Engn Res Ctr, Nanjing, Peoples R China
关键词
deep learning; neural network pruning; group sparsity; network compression;
D O I
10.1109/siprocess.2019.8868650
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Network pruning is important for the deployment of deep neural network on hardware platforms. However, most pruning methods focus on coarse-grained pruning with high precision loss. In addition, many fine-grained pruning methods focus the pruning on the fully connected layer rather than the convolution layer. The group pruning technique is proposed, focusing only on CONV-layer that enables to keep the weight reduction rate consistent within each weight group. This helps solve some inefficiency problems including internal buffer misalignment and load imbalance after fine-grained pruning. In the pre-training, the strategy of group sparse regularization (GSR) and standardizing the weight distribution is added to alleviate the loss of precision under high sparsity. Finally, we used the MNIST and CIFAR-10 data sets to test the LeNet and VGG-16 for experiments. The reduction rates of convolution layer are 87.5% and 62.5% respectively in the range of 0.14% accuracy loss.
引用
收藏
页码:325 / 329
页数:5
相关论文
共 50 条
  • [21] Dirichlet Pruning for Neural Network Compression
    Adamczewski, Kamil
    Park, Mijung
    24TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS (AISTATS), 2021, 130
  • [22] A lightweight deep neural network model and its applications based on channel pruning and group vector quantization
    Huang, Mingzhong
    Liu, Yan
    Zhao, Lijie
    Wang, Guogang
    NEURAL COMPUTING & APPLICATIONS, 2023, 36 (10): : 5333 - 5346
  • [23] A lightweight deep neural network model and its applications based on channel pruning and group vector quantization
    Mingzhong Huang
    Yan Liu
    Lijie Zhao
    Guogang Wang
    Neural Computing and Applications, 2024, 36 : 5333 - 5346
  • [24] Identity-linked Group Channel Pruning for Deep Neural Networks
    Zhang, Chenxin
    Xu, Keqin
    Liu, Jie
    Kang, Liangyi
    Zhou, Zhiyang
    Ye, Dan
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [25] Group Pruning Using a Bounded-lp Norm for Group Gating and Regularization
    Mummadi, Chaithanya Kumar
    Genewein, Tim
    Zhang, Dan
    Brox, Thomas
    Fischer, Volker
    PATTERN RECOGNITION, DAGM GCPR 2019, 2019, 11824 : 139 - 155
  • [26] Structured Sparsity of Convolutional Neural Networks via Nonconvex Sparse Group Regularization
    Bui, Kevin
    Park, Fredrick
    Zhang, Shuai
    Qi, Yingyong
    Xin, Jack
    FRONTIERS IN APPLIED MATHEMATICS AND STATISTICS, 2021, 6
  • [27] Convergence analyses on sparse feedforward neural networks via group lasso regularization
    Wang, Jian
    Cai, Qingling
    Chang, Qingquan
    Zurada, Jacek M.
    INFORMATION SCIENCES, 2017, 381 : 250 - 269
  • [28] MobilePrune: Neural Network Compression via l0 Sparse Group Lasso on the Mobile System
    Shao, Yubo
    Zhao, Kaikai
    Cao, Zhiwen
    Peng, Zhehao
    Peng, Xingang
    Li, Pan
    Wang, Yijie
    Ma, Jianzhu
    SENSORS, 2022, 22 (11)
  • [29] Deep neural network compression through interpretability-based filter pruning
    Yao, Kaixuan
    Cao, Feilong
    Leung, Yee
    Liang, Jiye
    PATTERN RECOGNITION, 2021, 119
  • [30] Sparse synthesis regularization with deep neural networks
    Obmann, Daniel
    Schwab, Johannes
    Haltmeier, Markus
    2019 13TH INTERNATIONAL CONFERENCE ON SAMPLING THEORY AND APPLICATIONS (SAMPTA), 2019,