Accelerating Neural Network Training: A Brief Review

被引:3
|
作者
Nokhwal, Sahil [1 ]
Chilakalapudi, Priyanka [1 ]
Donekal, Preeti [1 ]
Nokhwal, Suman [2 ]
Pahune, Saurabh [3 ]
Chaudhary, Ankit [4 ]
机构
[1] Univ Memphis, Memphis, TN 38152 USA
[2] Intercontinental Exchange Inc, Pleasanton, CA USA
[3] Cardinal Hlth, Dublin, OH USA
[4] Jawaharlal Nehru Univ, New Delhi, India
来源
2024 8TH INTERNATIONAL CONFERENCE ON INTELLIGENT SYSTEMS, METAHEURISTICS & SWARM INTELLIGENCE, ISMSI 2024 | 2024年
关键词
Neural Network Training; Acceleration Techniques; Training Optimization; Deep Learning Speedup; Model Training Efficiency; Machine Learning Accelerators; Training Time Reduction; Optimization Strategies;
D O I
10.1145/3665065.3665071
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The process of training a deep neural network is characterized by significant time requirements and associated costs. Although researchers have made considerable progress in this area, further work is still required due to resource constraints. This study examines innovative approaches to expedite the training process of deep neural networks (DNN), with specific emphasis on three state-of-the-art models such as ResNet50, Vision Transformer (ViT), and EfficientNet. The research utilizes sophisticated methodologies, including Gradient Accumulation (GA), Automatic Mixed Precision (AMP), and Pin Memory (PM), in order to optimize performance and accelerate the training procedure. The study examines the effects of these methodologies on the DNN models discussed earlier, assessing their efficacy with regard to training rate and computational efficacy. The study showcases the efficacy of including GA as a strategic approach, resulting in a noteworthy decrease in the duration required for training. This enables the models to converge at a faster pace. The utilization of AMP enhances the speed of computations by taking advantage of the advantages offered by lower precision arithmetic while maintaining the correctness of the model. Furthermore, this study investigates the application of Pin Memory as a strategy to enhance the efficiency of data transmission between the central processing unit and the graphics processing unit, thereby offering a promising opportunity for enhancing overall performance. The experimental findings demonstrate that the combination of these sophisticated methodologies significantly accelerates the training of DNNs, offering vital insights for experts seeking to improve the effectiveness of deep learning processes.
引用
收藏
页码:31 / 35
页数:5
相关论文
共 50 条
  • [21] Quantum Neural Network States: A Brief Review of Methods and Applications
    Jia, Zhih-Ahn
    Yi, Biao
    Zhai, Rui
    Wu, Yu-Chun
    Guo, Guang-Can
    Guo, Guo-Ping
    ADVANCED QUANTUM TECHNOLOGIES, 2019, 2 (7-8)
  • [22] Accelerating the Construction of Neural Network Potential Energy Surfaces: A Fast Hybrid Training Algorithm
    Zhang, Yao-long
    Zhou, Xue-yao
    Jiang, Bin
    CHINESE JOURNAL OF CHEMICAL PHYSICS, 2017, 30 (06) : 727 - 734
  • [23] Accelerating Data-Parallel Neural Network Training with Weighted-Averaging Reparameterisation
    Ramroach, Sterling
    Joshi, Ajay
    PARALLEL PROCESSING LETTERS, 2021, 31 (02)
  • [24] Accelerating Neural Network Training using Arbitrary Precision Approximating Matrix Multiplication Algorithms
    Ballard, Grey
    Weissenberger, Jack
    Zhang, Luoping
    50TH INTERNATIONAL CONFERENCE ON PARALLEL PROCESSING WORKSHOP PROCEEDINGS - ICPP WORKSHOPS '21, 2021,
  • [25] A method of accelerating neural network learning
    Sotirov, S
    NEURAL PROCESSING LETTERS, 2005, 22 (02) : 163 - 169
  • [26] Survey on Accelerating Neural Network with Hardware
    Chen G.
    Ma S.
    Guo Y.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2019, 56 (02): : 240 - 253
  • [27] A Method of Accelerating Neural Network Learning
    Sotir Sotirov
    Neural Processing Letters, 2005, 22 : 163 - 169
  • [28] Accelerating Training of Physics Informed Neural Network for 1D PDEs with Hierarchical Matrices
    Dobija, Mateusz
    Paszynska, Anna
    Uriarte, Carlos
    Paszynski, Maciej
    COMPUTATIONAL SCIENCE, ICCS 2024, PT III, 2024, 14834 : 352 - 362
  • [29] Review of Convolutional Neural Network Optimization and Training in Image Processing
    Ren, Yong
    Cheng, Xuemin
    TENTH INTERNATIONAL SYMPOSIUM ON PRECISION ENGINEERING MEASUREMENTS AND INSTRUMENTATION, 2019, 11053
  • [30] DeepAbstract: Neural Network Abstraction for Accelerating Verification
    Ashok, Pranav
    Hashemi, Vahid
    Kretinsky, Jan
    Mohr, Stefanie
    AUTOMATED TECHNOLOGY FOR VERIFICATION AND ANALYSIS (ATVA 2020), 2020, 12302 : 92 - 107