Systematic Analysis of Low-Precision Training in Deep Neural Networks: Factors Influencing Matrix Computations

被引:0
|
作者
Shen, Ao [1 ]
Lai, Zhiquan [1 ]
Zhang, Lizhi [1 ]
机构
[1] Natl Univ Def Technol, Natl Key Lab Parallel & Distributed Comp, Changsha 410073, Peoples R China
来源
APPLIED SCIENCES-BASEL | 2024年 / 14卷 / 21期
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
low-precision training; quantization noise; matrix computations; training efficiency; accumulation; frequency; depth;
D O I
10.3390/app142110025
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
As Deep Neural Networks (DNNs) continue to increase in complexity, the computational demands of their training have become a significant bottleneck. Low-precision training has emerged as a crucial strategy, wherein full-precision values are quantized to lower precisions, reducing computational overhead while aiming to maintain model accuracy. While prior research has primarily focused on minimizing quantization noise and optimizing performance for specific models and tasks, a comprehensive understanding of the general principles governing low-precision computations across diverse DNN architectures has been lacking. In this paper, we address this gap by systematically analyzing the factors that influence low-precision matrix computations, which are fundamental to DNN training. We investigate three critical factors-accumulation in matrix calculations, the frequency of element usage, and the depth of matrices within the model-and their impact on low-precision training. Through controlled experiments on standard models, as well as customized experiments designed to isolate individual factors, we derive several key insights: layers with higher accumulation and matrices with lower usage frequencies demonstrate greater tolerance to low-precision noise, without significantly compromising the stability of model training. Additionally, while the depth of matrices influences the stability of matrix operations to some extent, it does not have a noticeable effect on the overall training outcomes. Our findings contribute to the development of generalizable principles for low-precision training, offering a systematic framework applicable across various DNN architectures. We provide empirical evidence supporting the strategic allocation of training bit-widths based on the analyzed factors, thereby enhancing the efficiency and effectiveness of DNN training in resource-constrained environments.
引用
收藏
页数:18
相关论文
共 50 条
  • [1] POSITNN: TRAINING DEEP NEURAL NETWORKS WITH MIXED LOW-PRECISION POSIT
    Raposo, Goncalo
    Tomas, Pedro
    Roma, Nuno
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 7908 - 7912
  • [2] Training Deep Neural Networks in Low-Precision with High Accuracy using FPGAs
    Fox, Sean
    Faraone, Julian
    Boland, David
    Vissers, Kees
    Leong, Philip H. W.
    2019 INTERNATIONAL CONFERENCE ON FIELD-PROGRAMMABLE TECHNOLOGY (ICFPT 2019), 2019, : 1 - 9
  • [3] Customizing Low-Precision Deep Neural Networks For FPGAs
    Faraone, Julian
    Gambardella, Giulio
    Boland, David
    Fraser, Nicholas
    Blott, Michaela
    Leong, Philip H. W.
    2018 28TH INTERNATIONAL CONFERENCE ON FIELD PROGRAMMABLE LOGIC AND APPLICATIONS (FPL), 2018, : 97 - 100
  • [4] Role of Synaptic Stochasticity in Training Low-Precision Neural Networks
    Baldassi, Carlo
    Gerace, Federica
    Kappen, Hilbert J.
    Lucibello, Carlo
    Saglietti, Luca
    Tartaglione, Enzo
    Zecchina, Riccardo
    PHYSICAL REVIEW LETTERS, 2018, 120 (26)
  • [5] Neural Architecture Search for Low-Precision Neural Networks
    Wu, Binyi
    Waschneck, Bernd
    Mayr, Christian
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2022, PT IV, 2022, 13532 : 743 - 755
  • [6] Simulate-the-hardware: Training Accurate Binarized Neural Networks for Low-Precision Neural Accelerators
    Li, Jiajun
    Wang, Ying
    Liu, Bosheng
    Han, Yinhe
    Li, Xiaowei
    24TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC 2019), 2019, : 323 - 328
  • [7] Dataflow Architecture Optimization for Low-Precision Neural Networks
    Fan Z.
    Wu X.
    Li W.
    Cao H.
    An X.
    Ye X.
    Fan D.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2023, 60 (01): : 43 - 58
  • [8] Learning Sparse Low-Precision Neural Networks With Learnable Regularization
    Choi, Yoojin
    El-Khamy, Mostafa
    Lee, Jungwon
    IEEE ACCESS, 2020, 8 : 96963 - 96974
  • [9] Better schedules for low precision training of deep neural networks
    Cameron R. Wolfe
    Anastasios Kyrillidis
    Machine Learning, 2024, 113 : 3569 - 3587
  • [10] Better schedules for low precision training of deep neural networks
    Wolfe, Cameron R.
    Kyrillidis, Anastasios
    MACHINE LEARNING, 2024, 113 (06) : 3569 - 3587