Systematic Analysis of Low-Precision Training in Deep Neural Networks: Factors Influencing Matrix Computations

被引:0
|
作者
Shen, Ao [1 ]
Lai, Zhiquan [1 ]
Zhang, Lizhi [1 ]
机构
[1] Natl Univ Def Technol, Natl Key Lab Parallel & Distributed Comp, Changsha 410073, Peoples R China
来源
APPLIED SCIENCES-BASEL | 2024年 / 14卷 / 21期
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
low-precision training; quantization noise; matrix computations; training efficiency; accumulation; frequency; depth;
D O I
10.3390/app142110025
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
As Deep Neural Networks (DNNs) continue to increase in complexity, the computational demands of their training have become a significant bottleneck. Low-precision training has emerged as a crucial strategy, wherein full-precision values are quantized to lower precisions, reducing computational overhead while aiming to maintain model accuracy. While prior research has primarily focused on minimizing quantization noise and optimizing performance for specific models and tasks, a comprehensive understanding of the general principles governing low-precision computations across diverse DNN architectures has been lacking. In this paper, we address this gap by systematically analyzing the factors that influence low-precision matrix computations, which are fundamental to DNN training. We investigate three critical factors-accumulation in matrix calculations, the frequency of element usage, and the depth of matrices within the model-and their impact on low-precision training. Through controlled experiments on standard models, as well as customized experiments designed to isolate individual factors, we derive several key insights: layers with higher accumulation and matrices with lower usage frequencies demonstrate greater tolerance to low-precision noise, without significantly compromising the stability of model training. Additionally, while the depth of matrices influences the stability of matrix operations to some extent, it does not have a noticeable effect on the overall training outcomes. Our findings contribute to the development of generalizable principles for low-precision training, offering a systematic framework applicable across various DNN architectures. We provide empirical evidence supporting the strategic allocation of training bit-widths based on the analyzed factors, thereby enhancing the efficiency and effectiveness of DNN training in resource-constrained environments.
引用
收藏
页数:18
相关论文
共 50 条
  • [21] FPGA Prototyping of Low-Precision Zero-Skipping Accelerator for Neural Networks
    Kim, Dongyoung
    Kim, Soobeom
    Yoo, Sungjoo
    PROCEEDINGS OF THE 2018 29TH INTERNATIONAL SYMPOSIUM ON RAPID SYSTEM PROTOTYPING (RSP): SHORTENING THE PATH FROM SPECIFICATION TO PROTOTYPE, 2018, : 104 - 110
  • [22] IMPACT OF LOW-PRECISION DEEP REGRESSION NETWORKS ON SINGLE-CHANNEL SOURCE SEPARATION
    Ceolini, Enea
    Liu, Shih-Chii
    2017 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2017, : 256 - 260
  • [23] Ultra-Low Precision 4-bit Training of Deep Neural Networks
    Sun, Xiao
    Wang, Naigang
    Chen, Chia-Yu
    Ni, Jia-Min
    Agrawal, Ankur
    Cui, Xiaodong
    Venkataramani, Swagath
    El Maghraoui, Kaoutar
    Srinivasan, Vijayalakshmi
    Gopalakrishnan, Kailash
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [24] Learning Low-precision Neural Networks without Straight-Through Estimator (STE)
    Liu, Zhi-Gang
    Mattina, Matthew
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 3066 - 3072
  • [25] Quantized Neural Networks: Training Neural Networks with Low Precision Weights and Activations
    Hubara, Itay
    Courbariaux, Matthieu
    Soudry, Daniel
    El-Yaniv, Ran
    Bengio, Yoshua
    JOURNAL OF MACHINE LEARNING RESEARCH, 2018, 18
  • [26] CoNLoCNN: Exploiting Correlation and Non-Uniform Quantization for Energy-Efficient Low-precision Deep Convolutional Neural Networks
    Hanif, Muhammad Abdullah
    Sarda, Giuseppe Maria
    Marchisio, Alberto
    Masera, Guido
    Martina, Maurizio
    Shafique, Muhammad
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [27] SPECIAL ASPECTS OF MATRIX OPERATION IMPLEMENTATIONS FOR LOW-PRECISION NEURAL NETWORK MODEL ON THE ELBRUS PLATFORM
    Limonova, E. E.
    Neiman-zade, M., I
    Arlazarov, V. L.
    BULLETIN OF THE SOUTH URAL STATE UNIVERSITY SERIES-MATHEMATICAL MODELLING PROGRAMMING & COMPUTER SOFTWARE, 2020, 13 (01): : 118 - 128
  • [28] Training Deep Neural Networks with Low Precision Input Data: A Hurricane Prediction Case Study
    Kahira, Albert
    Bautista Gomez, Leonardo
    Badia, Rosa M.
    HIGH PERFORMANCE COMPUTING, ISC HIGH PERFORMANCE 2018, 2018, 11203 : 562 - 569
  • [29] LANCE: EFFICIENT LOW-PRECISION QUANTIZED WINOGRAD CONVOLUTION FOR NEURAL NETWORKS BASED ON GRAPHICS PROCESSING UNITS
    Li, Guangli
    Liu, Lei
    Wang, Xueying
    Ma, Xiu
    Feng, Xiaobing
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 3842 - 3846
  • [30] Systematic Measurement of Temperature Errors of Positive and Negative FOG Scale Factors Using a Low-Precision Turntable
    Pan, Jianye
    Zhou, Guofeng
    Li, Baoyu
    Gao, Feng
    Ma, Aojia
    Sun, Xiangchun
    IEEE SENSORS JOURNAL, 2023, 23 (14) : 16136 - 16144