Pruning-and-distillation: One-stage joint compression framework for CNNs via clustering

被引:4
|
作者
Niu, Tao [1 ]
Teng, Yinglei [1 ]
Jin, Lei [1 ]
Zou, Panpan [1 ]
Liu, Yiding [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Filter pruning; Clustering; Knowledge distillation; Deep neural networks; NEURAL-NETWORKS;
D O I
10.1016/j.imavis.2023.104743
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Network pruning and knowledge distillation, as two effective network compression techniques, have drawn extensive attention due to their success in reducing model complexity. However, previous works regard them as two independent methods and combine them in an isolated manner rather than joint, leading to a sub-optimal optimization. In this paper, we propose a collaborative compression scheme named Pruningand-Distillation via Clustering (PDC), which integrates pruning and distillation into an end-to-end single-stage framework that takes both advantages of them. Specifically, instead of directly deleting or zeroing out unimportant filters within each layer, we reconstruct them based on clustering, which preserves the learned features as much as possible. The guidance from the teacher is integrated into the pruning process to further improve the generalization of pruned model, which alleviates the randomness caused by reconstruction to some extent. After convergence, we can equivalently remove reconstructed filters within each cluster through the proposed channel addition operation. Benefiting from such equivalence, we no longer require the time-consuming finetuning step to regain accuracy. Extensive experiments on CIFAR-10/100 and ImageNet datasets show that our method achieves the best trade-off between performance and complexity compared with other state-of-theart algorithms. For example, for ResNet-110, we achieve a 61.5% FLOPs reduction with even 0.14% top-1 accuracy increase on CIFAR-10 and remove 55.2% FLOPs with only 0.32% accuracy drop on CIFAR-100. & COPY; 2023 Elsevier B.V. All rights reserved.
引用
收藏
页数:11
相关论文
共 50 条
  • [21] Effective Model Compression via Stage-wise Pruning
    Ming-Yang Zhang
    Xin-Yi Yu
    Lin-Lin Ou
    Machine Intelligence Research, 2023, 20 : 937 - 951
  • [22] Model compression via pruning and knowledge distillation for person re-identification
    Xie, Haonan
    Jiang, Wei
    Luo, Hao
    Yu, Hongyan
    JOURNAL OF AMBIENT INTELLIGENCE AND HUMANIZED COMPUTING, 2021, 12 (02) : 2149 - 2161
  • [23] Effective Model Compression via Stage-wise Pruning
    Zhang, Ming-Yang
    Yu, Xin-Yi
    Ou, Lin-Lin
    MACHINE INTELLIGENCE RESEARCH, 2023, 20 (06) : 937 - 951
  • [24] Model compression via pruning and knowledge distillation for person re-identification
    Haonan Xie
    Wei Jiang
    Hao Luo
    Hongyan Yu
    Journal of Ambient Intelligence and Humanized Computing, 2021, 12 : 2149 - 2161
  • [26] ONE-STAGE RECONSTRUCTION OF THE TEMPOROMANDIBULAR-JOINT IN HEMIFACIAL MICROSOMIA
    MUNRO, IR
    CHIR, B
    PLASTIC AND RECONSTRUCTIVE SURGERY, 1980, 66 (05) : 699 - 710
  • [27] Surgical Management of Periprosthetic Joint Infection: One-Stage Exchange
    Kendoff, Daniel
    Gehrke, Thorsten
    JOURNAL OF KNEE SURGERY, 2014, 27 (04) : 273 - 278
  • [28] One-stage Total Joint Arthroplasty for Patients With Active Tuberculosis
    Zhang, Yi-chao
    Zhang, Hong
    ORTHOPEDICS, 2013, 36 (05) : 328 - 330
  • [29] Lightweight One-Stage Maize Leaf Disease Detection Model with Knowledge Distillation
    Hu, Yanxin
    Liu, Gang
    Chen, Zhiyu
    Liu, Jiaqi
    Guo, Jianwei
    AGRICULTURE-BASEL, 2023, 13 (09):
  • [30] Exothermic isomerization in one-stage reactive distillation: Steady-state behaviour
    Vos, Floris Sebastiaan
    Bildea, Costin Sorin
    INDUSTRIAL & ENGINEERING CHEMISTRY RESEARCH, 2007, 46 (01) : 203 - 210