Accelerating Neural Network Training: A Brief Review

被引:3
|
作者
Nokhwal, Sahil [1 ]
Chilakalapudi, Priyanka [1 ]
Donekal, Preeti [1 ]
Nokhwal, Suman [2 ]
Pahune, Saurabh [3 ]
Chaudhary, Ankit [4 ]
机构
[1] Univ Memphis, Memphis, TN 38152 USA
[2] Intercontinental Exchange Inc, Pleasanton, CA USA
[3] Cardinal Hlth, Dublin, OH USA
[4] Jawaharlal Nehru Univ, New Delhi, India
来源
2024 8TH INTERNATIONAL CONFERENCE ON INTELLIGENT SYSTEMS, METAHEURISTICS & SWARM INTELLIGENCE, ISMSI 2024 | 2024年
关键词
Neural Network Training; Acceleration Techniques; Training Optimization; Deep Learning Speedup; Model Training Efficiency; Machine Learning Accelerators; Training Time Reduction; Optimization Strategies;
D O I
10.1145/3665065.3665071
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The process of training a deep neural network is characterized by significant time requirements and associated costs. Although researchers have made considerable progress in this area, further work is still required due to resource constraints. This study examines innovative approaches to expedite the training process of deep neural networks (DNN), with specific emphasis on three state-of-the-art models such as ResNet50, Vision Transformer (ViT), and EfficientNet. The research utilizes sophisticated methodologies, including Gradient Accumulation (GA), Automatic Mixed Precision (AMP), and Pin Memory (PM), in order to optimize performance and accelerate the training procedure. The study examines the effects of these methodologies on the DNN models discussed earlier, assessing their efficacy with regard to training rate and computational efficacy. The study showcases the efficacy of including GA as a strategic approach, resulting in a noteworthy decrease in the duration required for training. This enables the models to converge at a faster pace. The utilization of AMP enhances the speed of computations by taking advantage of the advantages offered by lower precision arithmetic while maintaining the correctness of the model. Furthermore, this study investigates the application of Pin Memory as a strategy to enhance the efficiency of data transmission between the central processing unit and the graphics processing unit, thereby offering a promising opportunity for enhancing overall performance. The experimental findings demonstrate that the combination of these sophisticated methodologies significantly accelerates the training of DNNs, offering vital insights for experts seeking to improve the effectiveness of deep learning processes.
引用
收藏
页码:31 / 35
页数:5
相关论文
共 50 条
  • [11] Accelerating neural network training with distributed asynchronous and selective optimization (DASO)
    Daniel Coquelin
    Charlotte Debus
    Markus Götz
    Fabrice von der Lehr
    James Kahn
    Martin Siggel
    Achim Streit
    Journal of Big Data, 9
  • [12] Accelerating distributed deep neural network training with pipelined MPI allreduce
    Castello, Adrian
    Quintana-Orti, Enrique S.
    Duato, Jose
    CLUSTER COMPUTING-THE JOURNAL OF NETWORKS SOFTWARE TOOLS AND APPLICATIONS, 2021, 24 (04): : 3797 - 3813
  • [13] Accelerating distributed deep neural network training with pipelined MPI allreduce
    Adrián Castelló
    Enrique S. Quintana-Ortí
    José Duato
    Cluster Computing, 2021, 24 : 3797 - 3813
  • [14] A Brief Review Over Neural Network Modeling Techniques
    Mohammadi, Mohammad Reza
    Sadrossadat, Sayed Alireza
    Mortazavi, Mir Gholamreza
    Noun, Behzad
    2017 IEEE INTERNATIONAL CONFERENCE ON POWER, CONTROL, SIGNALS AND INSTRUMENTATION ENGINEERING (ICPCSI), 2017, : 54 - 57
  • [15] The Neural Communication Network: A Brief Review on Neuromodeling and Neuroengineering
    Veletic, Mladen
    Balasingham, Ilangko
    2016 24TH TELECOMMUNICATIONS FORUM (TELFOR), 2016, : 74 - 77
  • [16] Accelerating Deep Neural Network training for autonomous landing guidance via homotopy
    Ni, Yang
    Pan, Binfeng
    Perez, Pablo Gomez
    ACTA ASTRONAUTICA, 2023, 212 : 654 - 664
  • [17] Accelerating Large-Scale Graph Neural Network Training on Crossbar Diet
    Ogbogu, Chukwufumnanya
    Arka, Aqeeb Iqbal
    Joardar, Biresh Kumar
    Doppa, Janardhan Rao
    Li, Hai
    Chakrabarty, Krishnendu
    Pande, Partha Pratim
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2022, 41 (11) : 3626 - 3637
  • [18] ACCELERATING RECURRENT NEURAL NETWORK TRAINING VIA TWO STAGE CLASSES AND PARALLELIZATION
    Huang, Zhiheng
    Zweig, Geoffrey
    Levit, Michael
    Dumoulin, Benoit
    Oguz, Barlas
    Chang, Shawn
    2013 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING (ASRU), 2013, : 326 - 331
  • [19] BulletTrain: Accelerating Robust Neural Network Training via Boundary Example Mining
    Hua, Weizhe
    Zhang, Yichi
    Guo, Chuan
    Zhang, Zhiru
    Suh, G. Edward
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [20] Accelerating Large-Scale Distributed Neural Network Training with SPMD Parallelism
    Zhang, Shiwei
    Diao, Lansong
    Wu, Chuan
    Wang, Siyu
    Lin, Wei
    PROCEEDINGS OF THE 13TH SYMPOSIUM ON CLOUD COMPUTING, SOCC 2022, 2022, : 403 - 418