Systematic Analysis of Low-Precision Training in Deep Neural Networks: Factors Influencing Matrix Computations

被引:0
|
作者
Shen, Ao [1 ]
Lai, Zhiquan [1 ]
Zhang, Lizhi [1 ]
机构
[1] Natl Univ Def Technol, Natl Key Lab Parallel & Distributed Comp, Changsha 410073, Peoples R China
来源
APPLIED SCIENCES-BASEL | 2024年 / 14卷 / 21期
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
low-precision training; quantization noise; matrix computations; training efficiency; accumulation; frequency; depth;
D O I
10.3390/app142110025
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
As Deep Neural Networks (DNNs) continue to increase in complexity, the computational demands of their training have become a significant bottleneck. Low-precision training has emerged as a crucial strategy, wherein full-precision values are quantized to lower precisions, reducing computational overhead while aiming to maintain model accuracy. While prior research has primarily focused on minimizing quantization noise and optimizing performance for specific models and tasks, a comprehensive understanding of the general principles governing low-precision computations across diverse DNN architectures has been lacking. In this paper, we address this gap by systematically analyzing the factors that influence low-precision matrix computations, which are fundamental to DNN training. We investigate three critical factors-accumulation in matrix calculations, the frequency of element usage, and the depth of matrices within the model-and their impact on low-precision training. Through controlled experiments on standard models, as well as customized experiments designed to isolate individual factors, we derive several key insights: layers with higher accumulation and matrices with lower usage frequencies demonstrate greater tolerance to low-precision noise, without significantly compromising the stability of model training. Additionally, while the depth of matrices influences the stability of matrix operations to some extent, it does not have a noticeable effect on the overall training outcomes. Our findings contribute to the development of generalizable principles for low-precision training, offering a systematic framework applicable across various DNN architectures. We provide empirical evidence supporting the strategic allocation of training bit-widths based on the analyzed factors, thereby enhancing the efficiency and effectiveness of DNN training in resource-constrained environments.
引用
收藏
页数:18
相关论文
共 50 条
  • [31] Training Neural Networks with Low Precision Dynamic Fixed-Point
    Jo, Sujeong
    Park, Hanmin
    Lee, Gunhee
    Choi, Kiyoung
    2018 IEEE 36TH INTERNATIONAL CONFERENCE ON COMPUTER DESIGN (ICCD), 2018, : 405 - 408
  • [32] Quantization-aware training for low precision photonic neural networks
    Kirtas, M.
    Oikonomou, A.
    Passalis, N.
    Mourgias-Alexandris, G.
    Moralis-Pegios, M.
    Pleros, N.
    Tefas, A.
    NEURAL NETWORKS, 2022, 155 : 561 - 573
  • [33] Training-aware Low Precision Quantization in Spiking Neural Networks
    Shymyrbay, Ayan
    Fouda, Mohammed E.
    Eltawil, Ahmed
    2022 56TH ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS, AND COMPUTERS, 2022, : 1147 - 1151
  • [34] Approximate Fisher Information Matrix to Characterize the Training of Deep Neural Networks
    Liao, Zhibin
    Drummond, Tom
    Reid, Ian
    Carneiro, Gustavo
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2020, 42 (01) : 15 - 26
  • [35] Impact of Mixed Precision Techniques on Training and Inference Efficiency of Deep Neural Networks
    Doerrich, Marion
    Fan, Mingcheng
    Kist, Andreas M.
    IEEE ACCESS, 2023, 11 : 57627 - 57634
  • [36] Full error analysis for the training of deep neural networks
    Beck, Christian
    Jentzen, Arnulf
    Kuckuck, Benno
    INFINITE DIMENSIONAL ANALYSIS QUANTUM PROBABILITY AND RELATED TOPICS, 2022, 25 (02)
  • [37] Mixed-precision architecture based on computational memory for training deep neural networks
    Nandakumar, S. R.
    Le Gallo, Manuel
    Boybat, Irem
    Rajendran, Bipin
    Sebastian, Abu
    Eleftheriou, Evangelos
    2018 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2018,
  • [38] Analysis of Deep Neural Networks with the Extended Data Jacobian Matrix
    Wang, Shengjie
    Mohamed, Abdel-rahman
    Caruana, Rich
    Bilmes, Jeff
    Plilipose, Matthai
    Richardson, Matthew
    Geras, Krzysztof
    Urban, Gregor
    Aslan, Ozlem
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48
  • [39] Sequence Discriminative Training for Low-Rank Deep Neural Networks
    Tachioka, Yuuki
    Watanabe, Shinji
    Le Roux, Jonathan
    Hershey, John R.
    2014 IEEE GLOBAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (GLOBALSIP), 2014, : 572 - 576
  • [40] Towards high performance low bitwidth training for deep neural networks
    Su, Chunyou
    Zhou, Sheng
    Feng, Liang
    Zhang, Wei
    JOURNAL OF SEMICONDUCTORS, 2020, 41 (02)