Facto-CNN: Memory-Efficient CNN Training with Low-rank Tensor Factorization and Lossy Tensor Compression

被引:0
|
作者
Lee, Seungtae [1 ]
Ko, Jonghwan [1 ]
Hong, Seokin [1 ]
机构
[1] Sungkyunkwan Univ, Suwon, South Korea
基金
新加坡国家研究基金会;
关键词
CNN; Training Optimization; Compression; Tensor Factorization;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Convolutional neural networks (CNNs) are becoming deeper and wider to achieve higher accuracy and lower loss, significantly expanding the computational resources. Especially, training CNN models extensively consumes memory mainly due to storing intermediate feature maps generated in the forward-propagation for calculating the gradient in the back-propagation. The memory usage of the CNN model training escalates with the increase in batch size and the complexity of the model. Therefore, a lightweight training method is essential, especially when the computational resources are limited. In this paper, we propose a CNN training mechanism called Facto-CNN, leveraging low-rank tensor factorization and lossy tensor compression to reduce the memory usage required in training the CNN models. Facto-CNN factorizes the weight tensors of convolutional and fully-connected layers and then only updates one of the factorized tensors for each layer, dramatically reducing the feature map size stored in the memory. To further reduce memory consumption, Facto-CNN compresses the feature maps with a simple lossy compression technique that exploits the value similarity in the feature maps. Our experimental evaluation demonstrates that Facto-CNN reduces the memory usage for storing the feature maps by 68-93% with a trivial accuracy degradation when training the CNN models.
引用
收藏
页数:16
相关论文
共 50 条
  • [31] Logarithmic Norm Regularized Low-Rank Factorization for Matrix and Tensor Completion
    Chen, Lin
    Jiang, Xue
    Liu, Xingzhao
    Zhou, Zhixin
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 3434 - 3449
  • [32] Imbalanced low-rank tensor completion via latent matrix factorization
    Qiu, Yuning
    Zhou, Guoxu
    Zeng, Junhua
    Zhao, Qibin
    Xie, Shengli
    NEURAL NETWORKS, 2022, 155 : 369 - 382
  • [33] Tensor completion using total variation and low-rank matrix factorization
    Ji, Teng-Yu
    Huang, Ting-Zhu
    Zhao, Xi-Le
    Ma, Tian-Hui
    Liu, Gang
    INFORMATION SCIENCES, 2016, 326 : 243 - 257
  • [34] Sparse and Low-Rank Constrained Tensor Factorization for Hyperspectral Image Unmixing
    Zheng, Pan
    Su, Hongjun
    Du, Qian
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2021, 14 (14) : 1754 - 1767
  • [35] Fast Nonnegative Matrix/Tensor Factorization Based on Low-Rank Approximation
    Zhou, Guoxu
    Cichocki, Andrzej
    Xie, Shengli
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2012, 60 (06) : 2928 - 2940
  • [36] Matrix factorization for low-rank tensor completion using framelet prior
    Jiang, Tai-Xiang
    Huang, Ting-Zhu
    Zhao, Xi-Le
    Ji, Teng-Yu
    Deng, Liang-Jian
    INFORMATION SCIENCES, 2018, 436 : 403 - 417
  • [37] Memory-Efficient CNN Accelerator Based on Interlayer Feature Map Compression
    Shao, Zhuang
    Chen, Xiaoliang
    Du, Li
    Chen, Lei
    Du, Yuan
    Zhuang, Wei
    Wei, Huadong
    Xie, Chenjia
    Wang, Zhongfeng
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2022, 69 (02) : 668 - 681
  • [38] A Low-Rank Tensor Factorization Using Implicit Similarity in Trust Relationships
    Ma, Pei
    Wang, Liejun
    Qin, Jiwei
    SYMMETRY-BASEL, 2020, 12 (03):
  • [39] Efficient Tensor Completion for Color Image and Video Recovery: Low-Rank Tensor Train
    Bengua, Johann A.
    Phien, Ho N.
    Hoang Duong Tuan
    Do, Minh N.
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2017, 26 (05) : 2466 - 2479
  • [40] Holistic CNN Compression via Low-Rank Decomposition with Knowledge Transfer
    Lin, Shaohui
    Ji, Rongrong
    Chen, Chao
    Tao, Dacheng
    Luo, Jiebo
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2019, 41 (12) : 2889 - 2905