Coarse-Grained Pruning of Neural Network Models Based on Blocky Sparse Structure

被引:1
|
作者
Huang, Lan [1 ,2 ]
Zeng, Jia [1 ]
Sun, Shiqi [1 ]
Wang, Wencong [1 ]
Wang, Yan [1 ,2 ]
Wang, Kangping [1 ,2 ]
机构
[1] Jilin Univ, Coll Comp Sci & Technol, Changchun 130012, Peoples R China
[2] Jilin Univ, Key Lab Symbol Computat & Knowledge Engn, Minist Educ, Changchun 130012, Peoples R China
基金
中国国家自然科学基金;
关键词
neural network compression; block pruning; sparse matrix computation; SPMV;
D O I
10.3390/e23081042
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
Deep neural networks may achieve excellent performance in many research fields. However, many deep neural network models are over-parameterized. The computation of weight matrices often consumes a lot of time, which requires plenty of computing resources. In order to solve these problems, a novel block-based division method and a special coarse-grained block pruning strategy are proposed in this paper to simplify and compress the fully connected structure, and the pruned weight matrices with a blocky structure are then stored in the format of Block Sparse Row (BSR) to accelerate the calculation of the weight matrices. First, the weight matrices are divided into square sub-blocks based on spatial aggregation. Second, a coarse-grained block pruning procedure is utilized to scale down the model parameters. Finally, the BSR storage format, which is much more friendly to block sparse matrix storage and computation, is employed to store these pruned dense weight blocks to speed up the calculation. In the following experiments on MNIST and Fashion-MNIST datasets, the trend of accuracies with different pruning granularities and different sparsity is explored in order to analyze our method. The experimental results show that our coarse-grained block pruning method can compress the network and can reduce the computational cost without greatly degrading the classification accuracy. The experiment on the CIFAR-10 dataset shows that our block pruning strategy can combine well with the convolutional networks.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Acceleration of Sparse Convolutional Neural Network Based on Coarse-Grained Dataflow Architecture
    Wu X.
    Ou Y.
    Li W.
    Wang D.
    Zhang H.
    Fan D.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2021, 58 (07): : 1504 - 1517
  • [2] Graph neural network based coarse-grained mapping prediction
    Li, Zhiheng
    Wellawatte, Geemi P.
    Chakraborty, Maghesree
    Gandhi, Heta A.
    Xu, Chenliang
    White, Andrew D.
    CHEMICAL SCIENCE, 2020, 11 (35) : 9524 - 9531
  • [3] COARSE-GRAINED DESCRIPTIONS OF OSCILLATIONS IN NEURONAL NETWORK MODELS
    Crodelle, Jennifer
    Newhall, Katherine A.
    Pyzza, Pamela B.
    Kovacic, Gregor
    COMMUNICATIONS IN MATHEMATICAL SCIENCES, 2019, 17 (05) : 1437 - 1458
  • [4] Coarse-grained models of water
    Darre, Leonardo
    Machado, Matias R.
    Pantano, Sergio
    WILEY INTERDISCIPLINARY REVIEWS-COMPUTATIONAL MOLECULAR SCIENCE, 2012, 2 (06) : 921 - 930
  • [5] Coarse-grained models for macromolecules
    Yethiraj, Arun
    ABSTRACTS OF PAPERS OF THE AMERICAN CHEMICAL SOCIETY, 2013, 246
  • [6] Coarse-grained models for proteins
    Tozzini, V
    CURRENT OPINION IN STRUCTURAL BIOLOGY, 2005, 15 (02) : 144 - 150
  • [7] Active learning a coarse-grained neural network model for bulk water from sparse training data
    Loeffler, Troy D.
    Patra, Tarak K.
    Chan, Henry
    Sankaranarayanan, Subramanian K. R. S.
    MOLECULAR SYSTEMS DESIGN & ENGINEERING, 2020, 5 (05) : 902 - 910
  • [8] A coarse-grained deep neural network model for liquid water
    Patra, Tarak K.
    Loeffler, Troy D.
    Chan, Henry
    Cherukara, Mathew J.
    Narayanan, Badri
    Sankaranarayanan, Subramanian K. R. S.
    APPLIED PHYSICS LETTERS, 2019, 115 (19)
  • [9] Neural Network Based Prediction of Conformational Free Energies - A New Route toward Coarse-Grained Simulation Models
    Lemke, Tobias
    Peter, Christine
    JOURNAL OF CHEMICAL THEORY AND COMPUTATION, 2017, 13 (12) : 6213 - 6221
  • [10] Exploring Explicit Coarse-Grained Structure in Artificial Neural Networks
    杨析辞
    谢志远
    杨晓涛
    Chinese Physics Letters, 2023, 40 (02) : 10 - 24