Convolutional Neural Network Compression via Tensor-Train Decomposition on Permuted Weight Tensor with Automatic Rank Determination

被引:2
|
作者
Gabor, Mateusz [1 ]
Zdunek, Rafal [1 ]
机构
[1] Wroclaw Univ Sci & Technol, Fac Elect Photon & Microsyst, Wybrzeze Wyspianskiego 27, PL-50370 Wroclaw, Poland
来源
关键词
Neural network compression; Convolutional neural network; Tensor decomposition; Tensor train decomposition;
D O I
10.1007/978-3-031-08757-8_54
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Convolutional neural networks (CNNs) are among the most commonly investigated models in computer vision. Deep CNNs yield high computational performance, but their common issue is a large size. For solving this problem, it is necessary to find effective compression methods which can effectively reduce the size of the network, keeping the accuracy on a similar level. This study provides important insights into the field of CNNs compression, introducing a novel low-rank compression method based on tensor-train decomposition on a permuted kernel weight tensor with automatic rank determination. The proposed method is easy to implement, and it allows us to fine-tune neural networks from decomposed factors instead of learning them from scratch. The results of this study examined on various CNN architectures and two datasets demonstrated that the proposed method outperforms other CNNs compression methods with respect to parameter and FLOPS compression at a low drop in the classification accuracy.
引用
收藏
页码:654 / 667
页数:14
相关论文
共 49 条
  • [31] Convolutional Neural Network Compression via Dynamic Parameter Rank Pruning
    Sharma, Manish
    Heard, Jamison
    Saber, Eli
    Markopoulos, Panagiotis
    IEEE ACCESS, 2025, 13 : 18441 - 18456
  • [32] Toward Near-Real-Time Training With Semi-Random Deep Neural Networks and Tensor-Train Decomposition
    Syed, Humza
    Bryla, Ryan
    Majumder, Uttam
    Kudithipudi, Dhireesha
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2021, 14 (14) : 8171 - 8179
  • [33] TT-SNN: Tensor Train Decomposition for Efficient Spiking Neural Network Training
    Lee, Donghyun
    Yin, Ruokai
    Kim, Youngeun
    Moitra, Abhishek
    Li, Yuhang
    Panda, Priyadarshini
    2024 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION, DATE, 2024,
  • [34] Exploiting Low-Rank Tensor-Train Deep Neural Networks Based on Riemannian Gradient Descent With Illustrations of Speech Processing
    Qi, Jun
    Yang, Chao-Han Huck
    Chen, Pin-Yu
    Tejedor, Javier
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2023, 31 : 633 - 642
  • [35] Bayesian Tensor Completion and Decomposition with Automatic CP Rank Determination Using MGP Shrinkage Prior
    Takayama H.
    Zhao Q.
    Hontani H.
    Yokota T.
    SN Computer Science, 2022, 3 (3)
  • [36] A Bayesian tensor ring decomposition model with automatic rank determination for spatiotemporal traffic data imputation
    Liu, Mengxia
    Lyu, Hao
    Ge, Hongxia
    Cheng, Rongjun
    APPLIED MATHEMATICAL MODELLING, 2025, 137
  • [37] Towards Compact Neural Networks via End-to-End Training: A Bayesian Tensor Approach with Automatic Rank Determination*
    Hawkins, Cole
    Liu, Xing
    Zhang, Zheng
    SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2022, 4 (01): : 46 - 71
  • [38] PARS: Proxy-Based Automatic Rank Selection for Neural Network Compression via Low-Rank Weight Approximation
    Sobolev, Konstantin
    Ermilov, Dmitry
    Phan, Anh-Huy
    Cichocki, Andrzej
    MATHEMATICS, 2022, 10 (20)
  • [39] TEC-CNN: Toward Efficient Compressing of Convolutional Neural Nets with Low-rank Tensor Decomposition
    Wang, Yifan
    Feng, Liang
    Cai, Fenglin
    Li, Lusi
    Wu, Rui
    Li, Jie
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2025, 21 (02)
  • [40] Infrared Image Monitoring Data Compression of Power Distribution Network via Tensor Tucker Decomposition
    Zhao H.
    Feng J.
    Ma L.
    Dianwang Jishu/Power System Technology, 2021, 45 (04): : 1632 - 1639