DEEP LEARNING BASED METHOD FOR PRUNING DEEP NEURAL NETWORKS

被引:18
|
作者
Li, Lianqiang [1 ]
Zhu, Jie [1 ]
Sun, Ming-Ting [2 ]
机构
[1] Shanghai Jiao Tong Univ, Dept Elect Engn, Shanghai, Peoples R China
[2] Univ Washington, Dept Elect & Comp Engn, Seattle, WA 98195 USA
基金
中国国家自然科学基金;
关键词
Network pruning; filter-level; deep learning;
D O I
10.1109/ICMEW.2019.00-68
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
In order to implement Deep Neural Networks (DNNs) into mobile devices, network pruning has been widely explored for lightening the complexity of Deep Neural Networks in terms of computational cost and parameter-storage load. In this paper, we propose a novel filter-level pruning method which utilizes a deep learning method to pursue compact DNNs. Specifically, we use a DNN model to extract features from the filters at first. Then, we employ a clustering algorithm to force the extracted features roll into different groups. By mapping the clustering results to the filters, we get the "similarity" relationships among the filters. At last, we keep the filter which is closest to the centroid in each cluster, prune out the others, and retrain the pruned DNN model. Compared with previous methods that employ heuristic ways on filters directly or selecting shallow features from filters manually, our method takes advantages of the deep learning method which can represent the raw filters in a more precise way. Experimental results show that our method outperforms several state-of-the-art pruning methods with negligible accuracy loss.
引用
收藏
页码:312 / 317
页数:6
相关论文
共 50 条
  • [21] Structured Pruning of Deep Convolutional Neural Networks
    Anwar, Sajid
    Hwang, Kyuyeon
    Sung, Wonyong
    ACM JOURNAL ON EMERGING TECHNOLOGIES IN COMPUTING SYSTEMS, 2017, 13 (03)
  • [22] Activation Pruning of Deep Convolutional Neural Networks
    Ardakani, Arash
    Condo, Carlo
    Gross, Warren J.
    2017 IEEE GLOBAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (GLOBALSIP 2017), 2017, : 1325 - 1329
  • [23] Fast Convex Pruning of Deep Neural Networks
    Aghasi, Alireza
    Abdi, Afshin
    Romberg, Justin
    SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2020, 2 (01): : 158 - 188
  • [24] Sparsity in deep learning: Pruning and growth for efficient inference and training in neural networks
    Hoefler, Torsten
    Alistarh, Dan
    Ben-Nun, Tal
    Dryden, Nikoli
    Peste, Alexandra
    Journal of Machine Learning Research, 2021, 22
  • [25] EDropout: Energy-Based Dropout and Pruning of Deep Neural Networks
    Salehinejad, Hojjat
    Valaee, Shahrokh
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (10) : 5279 - 5292
  • [26] SFP: Similarity-based filter pruning for deep neural networks
    Li, Guoqing
    Li, Rengang
    Li, Tuo
    Shen, Chaoyao
    Zou, Xiaofeng
    Wang, Jiuyang
    Wang, Changhong
    Li, Nanjun
    INFORMATION SCIENCES, 2025, 689
  • [27] Sparsity in Deep Learning: Pruning and growth for efficient inference and training in neural networks
    Hoefler, Torsten
    Alistarh, Dan
    Ben-Nun, Tal
    Dryden, Nikoli
    Peste, Alexandra
    JOURNAL OF MACHINE LEARNING RESEARCH, 2021, 23
  • [28] Automatic Pruning Rate Derivation for Structured Pruning of Deep Neural Networks
    Sakai, Yasufumi
    Iwakawa, Akinori
    Tabaru, Tsuguchika
    Inoue, Atsuki
    Kawaguchi, Hiroshi
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 2561 - 2567
  • [29] Deep representation-based transfer learning for deep neural networks
    Yang, Tao
    Yu, Xia
    Ma, Ning
    Zhang, Yifu
    Li, Hongru
    KNOWLEDGE-BASED SYSTEMS, 2022, 253
  • [30] SSFP: A Structured Stripe-Filter Pruning Method for Deep Neural Networks
    Liu, Jingjing
    Huang, Lingjin
    Feng, Manlong
    Guo, Aiying
    2024 13TH INTERNATIONAL CONFERENCE ON COMMUNICATIONS, CIRCUITS AND SYSTEMS, ICCCAS 2024, 2024, : 80 - 84