Pruning Deep Neural Networks with l0-constrained Optimization

被引:1
|
作者
Phan, Dzung T. [1 ]
Nguyen, Lam M. [1 ]
Nguyen, Nam H. [1 ]
Kalagnanam, Jayant R. [1 ]
机构
[1] IBM Res, Thomas J Watson Res Ctr, Yorktown Hts, NY 10598 USA
关键词
D O I
10.1109/ICDM50108.2020.00152
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) give state-of-the-art accuracy in many tasks, but they can require large amounts of memory storage, energy consumption, and long inference times. Modern DNNs can have hundreds of million parameters, which make it difficult for DNNs to be deployed in some applications with low-resource environments. Pruning redundant connections without sacrificing accuracy is one of popular approaches to overcome these limitations. We propose two l(0)-constrained optimization models for pruning deep neural networks layer-by-layer. The first model is devoted to a general activation function, while the second one is specifically for a ReLU. We introduce an efficient cutting plane algorithm to solve the latter to optimality. Our experiments show that the proposed approach achieves competitive compression rates over several state-of-the-art baseline methods.
引用
收藏
页码:1214 / 1219
页数:6
相关论文
共 50 条
  • [41] FRACTIONAL STEP DISCRIMINANT PRUNING: A FILTER PRUNING FRAMEWORK FOR DEEP CONVOLUTIONAL NEURAL NETWORKS
    Gkalelis, Nikolaos
    Mezaris, Vasileios
    2020 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO WORKSHOPS (ICMEW), 2020,
  • [42] TRP: Trained Rank Pruning for Efficient Deep Neural Networks
    Xu, Yuhui
    Li, Yuxi
    Zhang, Shuai
    Wen, Wei
    Wang, Botao
    Qi, Yingyong
    Chen, Yiran
    Lin, Weiyao
    Xiong, Hongkai
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 977 - 983
  • [43] Effective Search Space Pruning for Testing Deep Neural Networks
    Rangayah, Bala
    Sng, Eugene
    Trinh, Minh-Thai
    PROGRAMMING LANGUAGES AND SYSTEMS, APLAS 2024, 2025, 15194 : 365 - 387
  • [44] Pruning Deep Convolutional Neural Networks Architectures with Evolution Strategy
    Fernandes, Francisco E., Jr.
    Yen, Gary G.
    INFORMATION SCIENCES, 2021, 552 : 29 - 47
  • [45] Can Unstructured Pruning Reduce the Depth in Deep Neural Networks?
    Liao, Zhu
    Quetu, Victor
    Nguyen, Van-Tam
    Tartaglione, Enzo
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 1394 - 1398
  • [46] Heuristic-based automatic pruning of deep neural networks
    Choudhary, Tejalal
    Mishra, Vipul
    Goswami, Anurag
    Sarangapani, Jagannathan
    NEURAL COMPUTING & APPLICATIONS, 2022, 34 (06): : 4889 - 4903
  • [47] Heuristic-based automatic pruning of deep neural networks
    Tejalal Choudhary
    Vipul Mishra
    Anurag Goswami
    Jagannathan Sarangapani
    Neural Computing and Applications, 2022, 34 : 4889 - 4903
  • [48] Class-Separation Preserving Pruning for Deep Neural Networks
    Preet I.
    Boydell O.
    John D.
    IEEE Transactions on Artificial Intelligence, 2024, 5 (01): : 290 - 299
  • [49] Gradient and Magnitude Based Pruning for Sparse Deep Neural Networks
    Belay, Kaleab
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 13126 - 13127
  • [50] Asymptotic Soft Filter Pruning for Deep Convolutional Neural Networks
    He, Yang
    Dong, Xuanyi
    Kang, Guoliang
    Fu, Yanwei
    Yan, Chenggang
    Yang, Yi
    IEEE TRANSACTIONS ON CYBERNETICS, 2020, 50 (08) : 3594 - 3604