Accelerating Convolutional Neural Networks with Dominant Convolutional Kernel and Knowledge Pre-regression

被引:13
|
作者
Wang, Zhenyang [1 ]
Deng, Zhidong [1 ]
Wang, Shiyao [1 ]
机构
[1] Tsinghua Univ, Dept Comp Sci, Tsinghua Natl Lab Informat Sci & Technol, State Key Lab Intelligent Technol & Syst, Beijing 100084, Peoples R China
来源
关键词
Dominant convolutional kernel; Knowledge pre-regression; Model compression; Knowledge distilling;
D O I
10.1007/978-3-319-46484-8_32
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Aiming at accelerating the test time of deep convolutional neural networks (CNNs), we propose a model compression method that contains a novel dominant kernel (DK) and a new training method called knowledge pre-regression (KP). In the combined model DK(2)PNet, DK is presented to significantly accomplish a low-rank decomposition of convolutional kernels, while KP is employed to transfer knowledge of intermediate hidden layers from a larger teacher network to its compressed student network on the basis of a cross entropy loss function instead of previous Euclidean distance. Compared to the latest results, the experimental results achieved on CIFAR-10, CIFAR-100, MNIST, and SVHN benchmarks show that our DK(2)PNet method has the best performance in the light of being close to the state of the art accuracy and requiring dramatically fewer number of model parameters.
引用
收藏
页码:533 / 548
页数:16
相关论文
共 50 条
  • [41] Knowledge-Driven Interpretation of Convolutional Neural Networks
    Massidda, Riccardo
    Bacciu, Davide
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2022, PT I, 2023, 13713 : 356 - 371
  • [42] Accelerating Binarized Convolutional Neural Networks with Software-Programmable FPGAs
    Zhao, Ritchie
    Song, Weinan
    Zhang, Wentao
    Xing, Tianwei
    Lin, Jeng-Hau
    Srivastava, Mani
    Gupta, Rajesh
    Zhang, Zhiru
    FPGA'17: PROCEEDINGS OF THE 2017 ACM/SIGDA INTERNATIONAL SYMPOSIUM ON FIELD-PROGRAMMABLE GATE ARRAYS, 2017, : 15 - 24
  • [43] RepSGD: Channel Pruning Using Reparamerization for Accelerating Convolutional Neural Networks
    Kim, Nam Joon
    Kim, Hyun
    2023 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS, ISCAS, 2023,
  • [44] SOURCE-SEPARATED AUDIO INPUT FOR ACCELERATING CONVOLUTIONAL NEURAL NETWORKS
    Dominguez, Miguel
    Daigneau, Madeleine
    Ptucha, Raymond
    2017 IEEE WESTERN NEW YORK IMAGE AND SIGNAL PROCESSING WORKSHOP (WNYISPW), 2017,
  • [45] Exploring Heterogeneous Algorithms for Accelerating Deep Convolutional Neural Networks on FPGAs
    Xiao, Qincheng
    Liang, Yun
    Lu, Liqiang
    Yan, Shengen
    Tai, Yu-Wing
    PROCEEDINGS OF THE 2017 54TH ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2017,
  • [46] Tutor-Instructing Global Pruning for Accelerating Convolutional Neural Networks
    Yu, Fang
    Cui, Li
    ECAI 2020: 24TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, 325 : 2792 - 2799
  • [47] Channel pruning based on mean gradient for accelerating Convolutional Neural Networks
    Liu, Congcong
    Wu, Huaming
    SIGNAL PROCESSING, 2019, 156 : 84 - 91
  • [48] Accelerating Deep Convolutional Neural Networks Using Number Theoretic Transform
    Hong, Seongmin
    Arthanto, Yashael Faith
    Kim, Joo-Young
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2023, 70 (01) : 315 - 326
  • [49] Accelerating Inference of Convolutional Neural Networks Using In-memory Computing
    Dazzi, Martino
    Sebastian, Abu
    Benini, Luca
    Eleftheriou, Evangelos
    FRONTIERS IN COMPUTATIONAL NEUROSCIENCE, 2021, 15
  • [50] Complex hybrid weighted pruning method for accelerating convolutional neural networks
    Geng, Xu
    Gao, Jinxiong
    Zhang, Yonghui
    Xu, Dingtan
    SCIENTIFIC REPORTS, 2024, 14 (01)