Layer-Wise Data-Free CNN Compression

被引:1
|
作者
Horton, Maxwell [1 ]
Jin, Yanzi [1 ]
Farhadi, Ali [1 ]
Rastegari, Mohammad [1 ]
机构
[1] Apple, Cupertino, CA 95014 USA
关键词
D O I
10.1109/ICPR56361.2022.9956237
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present a computationally efficient method for compressing a trained neural network without using real data. We break the problem of data-free network compression into independent layer-wise compressions. We show how to efficiently generate layer-wise training data using only a pretrained network. We use this data to perform independent layer-wise compressions on the pretrained network. We also show how to precondition the network to improve the accuracy of our layer-wise compression method. We present results for layer-wise compression using quantization and pruning. When quantizing, we compress with higher accuracy than related works while using orders of magnitude less compute. When compressing MobileNetV2 and evaluating on ImageNet, our method outperforms existing methods for quantization at all bit-widths, achieving a +0.34% improvement in 8-bit quantization, and a stronger improvement at lower bit-widths (up to a +28.50% improvement at 5 bits). When pruning, we outperform baselines of a similar compute envelope, achieving 1.5 times the sparsity rate at the same accuracy. We also show how to combine our efficient method with high-compute generative methods to improve upon their results.
引用
收藏
页码:2019 / 2026
页数:8
相关论文
共 50 条
  • [21] Source data-free domain adaptation for a faster R-CNN
    Xiong, Lin
    Ye, Mao
    Zhang, Dan
    Gan, Yan
    Liu, Yiguang
    PATTERN RECOGNITION, 2022, 124
  • [22] Dual discriminator adversarial distillation for data-free model compression
    Zhao, Haoran
    Sun, Xin
    Dong, Junyu
    Manic, Milos
    Zhou, Huiyu
    Yu, Hui
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2022, 13 (05) : 1213 - 1230
  • [23] Dual discriminator adversarial distillation for data-free model compression
    Haoran Zhao
    Xin Sun
    Junyu Dong
    Milos Manic
    Huiyu Zhou
    Hui Yu
    International Journal of Machine Learning and Cybernetics, 2022, 13 : 1213 - 1230
  • [24] Transkimmer: Transformer Learns to Layer-wise Skim
    Guan, Yue
    Li, Zhengyi
    Leng, Jingwen
    Lin, Zhouhan
    Guo, Minyi
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 7275 - 7286
  • [25] Layer-Wise Representation Fusion for Compositional Generalization
    Zheng, Yafang
    Lin, Lei
    Li, Shuangtao
    Yuan, Yuxuan
    Lai, Zhaohong
    Liu, Shan
    Fu, Biao
    Chen, Yidong
    Shi, Xiaodong
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 17, 2024, : 19706 - 19714
  • [26] MixMix: All You Need for Data-Free Compression Are Feature and Data Mixing
    Li, Yuhang
    Zhu, Feng
    Gong, Ruihao
    Shen, Mingzhu
    Dong, Xin
    Yu, Fengwei
    Lu, Shaoqing
    Gu, Shi
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 4390 - 4399
  • [27] FedLF: Layer-Wise Fair Federated Learning
    Pan, Zibin
    Li, Chi
    Yu, Fangchen
    Wang, Shuyi
    Wang, Haijin
    Tang, Xiaoying
    Zhao, Junhua
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 13, 2024, : 14527 - 14535
  • [28] SPSA for Layer-Wise Training of Deep Networks
    Wulff, Benjamin
    Schuecker, Jannis
    Bauckhage, Christian
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2018, PT III, 2018, 11141 : 564 - 573
  • [29] FLEXIBLE NETWORK BINARIZATION WITH LAYER-WISE PRIORITY
    Wang, He
    Xu, Yi
    Ni, Bingbing
    Zhuang, Lixue
    Xu, Hongteng
    2018 25TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2018, : 2346 - 2350
  • [30] Layer-Wise Personalized Federated Learning with Hypernetwork
    Suxia Zhu
    Tianyu Liu
    Guanglu Sun
    Neural Processing Letters, 2023, 55 (9) : 12273 - 12287