Layer-Wise Data-Free CNN Compression

被引:1
|
作者
Horton, Maxwell [1 ]
Jin, Yanzi [1 ]
Farhadi, Ali [1 ]
Rastegari, Mohammad [1 ]
机构
[1] Apple, Cupertino, CA 95014 USA
关键词
D O I
10.1109/ICPR56361.2022.9956237
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present a computationally efficient method for compressing a trained neural network without using real data. We break the problem of data-free network compression into independent layer-wise compressions. We show how to efficiently generate layer-wise training data using only a pretrained network. We use this data to perform independent layer-wise compressions on the pretrained network. We also show how to precondition the network to improve the accuracy of our layer-wise compression method. We present results for layer-wise compression using quantization and pruning. When quantizing, we compress with higher accuracy than related works while using orders of magnitude less compute. When compressing MobileNetV2 and evaluating on ImageNet, our method outperforms existing methods for quantization at all bit-widths, achieving a +0.34% improvement in 8-bit quantization, and a stronger improvement at lower bit-widths (up to a +28.50% improvement at 5 bits). When pruning, we outperform baselines of a similar compute envelope, achieving 1.5 times the sparsity rate at the same accuracy. We also show how to combine our efficient method with high-compute generative methods to improve upon their results.
引用
收藏
页码:2019 / 2026
页数:8
相关论文
共 50 条
  • [41] Learning Layer-wise Equivariances Automatically using Gradients
    van der Ouderaa, Tycho F. A.
    Immer, Alexander
    van der Wilk, Mark
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [42] MLP in layer-wise form with applications to weight decay
    Kärkkäinen, T
    NEURAL COMPUTATION, 2002, 14 (06) : 1451 - 1480
  • [43] A layer-wise frequency scaling for a neural processing unit
    Chung, Jaehoon
    Kim, HyunMi
    Shin, Kyoungseon
    Lyuh, Chun-Gi
    Cho, Yong Cheol Peter
    Han, Jinho
    Kwon, Youngsu
    Gong, Young-Ho
    Chung, Sung Woo
    ETRI JOURNAL, 2022, 44 (05) : 849 - 858
  • [44] Layer-wise domain correction for unsupervised domain adaptation
    Shuang Li
    Shi-ji Song
    Cheng Wu
    Frontiers of Information Technology & Electronic Engineering, 2018, 19 : 91 - 103
  • [45] Layer-Wise Weight Decay for Deep Neural Networks
    Ishii, Masato
    Sato, Atsushi
    IMAGE AND VIDEO TECHNOLOGY (PSIVT 2017), 2018, 10749 : 276 - 289
  • [46] Adversarial Examples Detection and Analysis with Layer-wise Autoencoders
    Wojcik, Bartosz
    Morawiecki, Pawel
    Smieja, Marek
    Krzyzek, Tomasz
    Spurek, Przemyslaw
    Tabor, Jacek
    2021 IEEE 33RD INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2021), 2021, : 1322 - 1326
  • [47] Layer-wise domain correction for unsupervised domain adaptation
    Shuang LI
    Shi-ji SONG
    Cheng WU
    FrontiersofInformationTechnology&ElectronicEngineering, 2018, 19 (01) : 91 - 103
  • [48] Layer-wise spatial modeling of porosity in additive manufacturing
    Liu, Jia
    Liu, Chenang
    Bai, Yun
    Rao, Prahalada
    Williams, Christopher B.
    Kong, Zhenyu
    IISE TRANSACTIONS, 2019, 51 (02) : 109 - 123
  • [49] Layer-Wise Discontinuous Galerkin Methods for Piezoelectric Laminates
    Benedetti, Ivano
    Gulizzi, Vincenzo
    Milazzo, Alberto
    MODELLING, 2020, 1 (02): : 198 - 214
  • [50] Towards layer-wise quantization for heterogeneous federated clients
    Xu, Yang
    Cheng, Junhao
    Xu, Hongli
    Guo, Changyu
    Liao, Yunming
    Yao, Zhiwei
    COMPUTER NETWORKS, 2025, 264