Convolutional Neural Network Compression via Tensor-Train Decomposition on Permuted Weight Tensor with Automatic Rank Determination

被引:2
|
作者
Gabor, Mateusz [1 ]
Zdunek, Rafal [1 ]
机构
[1] Wroclaw Univ Sci & Technol, Fac Elect Photon & Microsyst, Wybrzeze Wyspianskiego 27, PL-50370 Wroclaw, Poland
来源
关键词
Neural network compression; Convolutional neural network; Tensor decomposition; Tensor train decomposition;
D O I
10.1007/978-3-031-08757-8_54
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Convolutional neural networks (CNNs) are among the most commonly investigated models in computer vision. Deep CNNs yield high computational performance, but their common issue is a large size. For solving this problem, it is necessary to find effective compression methods which can effectively reduce the size of the network, keeping the accuracy on a similar level. This study provides important insights into the field of CNNs compression, introducing a novel low-rank compression method based on tensor-train decomposition on a permuted kernel weight tensor with automatic rank determination. The proposed method is easy to implement, and it allows us to fine-tune neural networks from decomposed factors instead of learning them from scratch. The results of this study examined on various CNN architectures and two datasets demonstrated that the proposed method outperforms other CNNs compression methods with respect to parameter and FLOPS compression at a low drop in the classification accuracy.
引用
收藏
页码:654 / 667
页数:14
相关论文
共 49 条
  • [21] TTH-RNN: Tensor-Train Hierarchical Recurrent Neural Network for Video Summarization
    Zhao, Bin
    Li, Xuelong
    Lu, Xiaoqiang
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2021, 68 (04) : 3629 - 3637
  • [22] Deep Convolutional Neural Network Compression Method: Tensor Ring Decomposition with Variational Bayesian Approach
    Liu, Weirong
    Zhang, Min
    Shi, Changhong
    Zhang, Ning
    Liu, Jie
    NEURAL PROCESSING LETTERS, 2024, 56 (02)
  • [23] Deep Convolutional Neural Network Compression Method: Tensor Ring Decomposition with Variational Bayesian Approach
    Weirong Liu
    Min Zhang
    Changhong Shi
    Ning Zhang
    Jie Liu
    Neural Processing Letters, 56
  • [24] Distributed Nonnegative Tensor Canonical Polyadic Decomposition With Automatic Rank Determination
    Cheng, Lei
    Tong, Xueke
    Wu, Yik-Chung
    2020 IEEE 11TH SENSOR ARRAY AND MULTICHANNEL SIGNAL PROCESSING WORKSHOP (SAM), 2020,
  • [25] Tensor Network Methods for Hyperparameter Optimization and Compression of Convolutional Neural Networks
    Naumov, A.
    Melnikov, A.
    Perelshtein, M.
    Melnikov, Ar.
    Abronin, V.
    Oksanichenko, F.
    APPLIED SCIENCES-BASEL, 2025, 15 (04):
  • [26] Convolutional Neural Network Feature Extraction Using Covariance Tensor Decomposition
    Fonseca, Ricardo
    Guarnizo, Oscar
    Suntaxi, Diego
    Cadiz, Alfonso
    Creixell, Werner
    IEEE ACCESS, 2021, 9 : 66646 - 66660
  • [27] CP-decomposition with Tensor Power Method for Convolutional Neural Networks Compression
    Astrid, Marcella
    Lee, Seung-Ik
    2017 IEEE INTERNATIONAL CONFERENCE ON BIG DATA AND SMART COMPUTING (BIGCOMP), 2017, : 115 - 118
  • [28] Recurrent Neural Network Compression Based on Low-Rank Tensor Representation
    Tjandra, Andros
    Sakti, Sakriani
    Nakamura, Satoshi
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2020, E103D (02) : 435 - 449
  • [29] A Low-Complexity Neural Normalized Min-Sum LDPC Decoding Algorithm Using Tensor-Train Decomposition
    Liang, Yuanhui
    Lam, Chan-Tong
    Ng, Benjamin K.
    IEEE COMMUNICATIONS LETTERS, 2022, 26 (12) : 2914 - 2918
  • [30] Hardware Acceleration in Large-Scale Tensor Decomposition for Neural Network Compression
    Kao, Chen-Chien
    Hsieh, Yi-Yen
    Chen, Chao-Hung
    Yang, Chia-Hsiang
    2022 IEEE 65TH INTERNATIONAL MIDWEST SYMPOSIUM ON CIRCUITS AND SYSTEMS (MWSCAS 2022), 2022,