Accelerating Convolutional Neural Networks with Dominant Convolutional Kernel and Knowledge Pre-regression

被引:13
|
作者
Wang, Zhenyang [1 ]
Deng, Zhidong [1 ]
Wang, Shiyao [1 ]
机构
[1] Tsinghua Univ, Dept Comp Sci, Tsinghua Natl Lab Informat Sci & Technol, State Key Lab Intelligent Technol & Syst, Beijing 100084, Peoples R China
来源
关键词
Dominant convolutional kernel; Knowledge pre-regression; Model compression; Knowledge distilling;
D O I
10.1007/978-3-319-46484-8_32
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Aiming at accelerating the test time of deep convolutional neural networks (CNNs), we propose a model compression method that contains a novel dominant kernel (DK) and a new training method called knowledge pre-regression (KP). In the combined model DK(2)PNet, DK is presented to significantly accomplish a low-rank decomposition of convolutional kernels, while KP is employed to transfer knowledge of intermediate hidden layers from a larger teacher network to its compressed student network on the basis of a cross entropy loss function instead of previous Euclidean distance. Compared to the latest results, the experimental results achieved on CIFAR-10, CIFAR-100, MNIST, and SVHN benchmarks show that our DK(2)PNet method has the best performance in the light of being close to the state of the art accuracy and requiring dramatically fewer number of model parameters.
引用
收藏
页码:533 / 548
页数:16
相关论文
共 50 条
  • [21] A Digitally Controlled Analog kernel for Convolutional Neural Networks
    Asghar, Malik Summair
    Junaid, Muhammad
    Kim, Hyung Won
    Arslan, Saad
    Shah, Syed Asmat Ali
    18TH INTERNATIONAL SOC DESIGN CONFERENCE 2021 (ISOCC 2021), 2021, : 242 - 243
  • [22] Efficient Weighted Kernel Sharing Convolutional Neural Networks
    Zhou, Helong
    Chen, Yie-Tarng
    Zhang, Jie
    Fang, Wen-Hsien
    2018 IEEE INTERNATIONAL CONFERENCE ON VISUAL COMMUNICATIONS AND IMAGE PROCESSING (IEEE VCIP), 2018,
  • [23] V-SKP: Vectorized Kernel-Based Structured Kernel Pruning for Accelerating Deep Convolutional Neural Networks
    Koo, Kwanghyun
    Kim, Hyun
    IEEE ACCESS, 2023, 11 : 118547 - 118557
  • [24] Accelerating Convergence of Fluid Dynamics Simulations with Convolutional Neural Networks
    Hajgato, Gergely
    Gyires-Toth, Balint
    Paal, Gyorgy
    PERIODICA POLYTECHNICA-MECHANICAL ENGINEERING, 2019, 63 (03): : 230 - 239
  • [25] Accelerating Convolutional Neural Networks via Activation Map Compression
    Georgiadis, Georgios
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 7078 - 7088
  • [26] Accelerating Sparse Convolutional Neural Networks Based on Dataflow Architecture
    Wu, Xinxin
    Li, Yi
    Ou, Yan
    Li, Wenming
    Sun, Shibo
    Xu, Wenxing
    Fan, Dongrui
    ALGORITHMS AND ARCHITECTURES FOR PARALLEL PROCESSING, ICA3PP 2020, PT II, 2020, 12453 : 14 - 31
  • [27] Soft Taylor Pruning for Accelerating Deep Convolutional Neural Networks
    Rong, Jintao
    Yu, Xiyi
    Zhang, Mingyang
    Ou, Linlin
    IECON 2020: THE 46TH ANNUAL CONFERENCE OF THE IEEE INDUSTRIAL ELECTRONICS SOCIETY, 2020, : 5343 - 5349
  • [28] Accelerating Convolutional Neural Networks by Exploiting the Sparsity of Output Activation
    Fan, Zhihua
    Li, Wenming
    Wang, Zhen
    Liu, Tianyu
    Wu, Haibin
    Liu, Yanhuan
    Wu, Meng
    Wu, Xinxin
    Ye, Xiaochun
    Fan, Dongrui
    Sun, Ninghui
    An, Xuejun
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2023, 34 (12) : 3253 - 3265
  • [29] Accelerating Brain MR Imaging With Multisequence and Convolutional Neural Networks
    Mo, Zhanhao
    Sui, He
    Lv, Zhongwen
    Huang, Xiaoqian
    Li, Guobin
    Shen, Dinggang
    Liu, Lin
    Liao, Shu
    BRAIN AND BEHAVIOR, 2024, 14 (11):
  • [30] A CGRA-based Approach for Accelerating Convolutional Neural Networks
    Tanomoto, Masakazu
    Takamaeda-Yamazaki, Shinya
    Yao, Jun
    Nakashima, Yasuhiko
    2015 IEEE 9TH INTERNATIONAL SYMPOSIUM ON EMBEDDED MULTICORE/MANYCORE SYSTEMS-ON-CHIP (MCSOC), 2015, : 73 - 80