Lightweight and Elegant Data Reduction Strategies for Training Acceleration of Convolutional Neural Networks

被引:0
|
作者
Demidovskij, Alexander [1 ,2 ]
Tugaryov, Artyom [1 ]
Trutnev, Aleksei [1 ]
Kazyulina, Marina [1 ]
Salnikov, Igor [1 ]
Pavlov, Stanislav [1 ]
机构
[1] Huawei Russian Res Inst, NN AI Team, Ul Maksima Gorkogo 117, Nizhnii Novgorod 603006, Russia
[2] Natl Res Univ, Higher Sch Econ, Dept Informat Math & Comp Sci, Ul Bolshaya Pecherskaya 25-12, Nizhnii Novgorod 603155, Russia
关键词
deep learning training; training acceleration; convolutional neural networks; sample importance; dataset reduction;
D O I
10.3390/math11143120
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Due to industrial demands to handle increasing amounts of training data, lower the cost of computing one model at a time, and lessen the ecological effects of intensive computing resource consumption, the job of speeding the training of deep neural networks becomes exceedingly challenging. Adaptive Online Importance Sampling and IDS are two brand-new methods for accelerating training that are presented in this research. On the one hand, Adaptive Online Importance Sampling accelerates neural network training by lowering the number of forward and backward steps depending on how poorly a model can identify a given data sample. On the other hand, Intellectual Data Selection accelerates training by removing semantic redundancies from the training dataset and subsequently lowering the number of training steps. The study reports average 1.9x training acceleration for ResNet50, ResNet18, MobileNet v2 and YOLO v5 on a variety of datasets: CIFAR-100, CIFAR-10, ImageNet 2012 and MS COCO 2017, where training data are reduced by up to five times. Application of Adaptive Online Importance Sampling to ResNet50 training on ImageNet 2012 results in 2.37 times quicker convergence to 71.7% top-1 accuracy, which is within 5% of the baseline. Total training time for the same number of epochs as the baseline is reduced by 1.82 times, with an accuracy drop of 2.45 p.p. The amount of time required to apply Intellectual Data Selection to ResNet50 training on ImageNet 2012 is decreased by 1.27 times with a corresponding decline in accuracy of 1.12 p.p. Applying both methods to ResNet50 training on ImageNet 2012 results in 2.31 speedup with an accuracy drop of 3.5 p.p.
引用
收藏
页数:25
相关论文
共 50 条
  • [1] Generative adversarial network based synthetic data training model for lightweight convolutional neural networks
    Rather, Ishfaq Hussain
    Kumar, Sushil
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 83 (2) : 6249 - 6271
  • [2] Generative adversarial network based synthetic data training model for lightweight convolutional neural networks
    Ishfaq Hussain Rather
    Sushil Kumar
    Multimedia Tools and Applications, 2024, 83 : 6249 - 6271
  • [3] Training Strategies for Convolutional Neural Networks with Transformed Input
    Khandani, Masoumeh Kalantari
    Mikhael, Wasfy B.
    2021 IEEE INTERNATIONAL MIDWEST SYMPOSIUM ON CIRCUITS AND SYSTEMS (MWSCAS), 2021, : 1058 - 1061
  • [4] CONVOLUTIONAL NEURAL NETWORKS AND TRAINING STRATEGIES FOR SKIN DETECTION
    Kim, Yoonsik
    Hwang, Insung
    Cho, Nam Ik
    2017 24TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2017, : 3919 - 3923
  • [5] Data Dropout: Optimizing Training Data for Convolutional Neural Networks
    Wang, Tianyang
    Huan, Jun
    Li, Bo
    2018 IEEE 30TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI), 2018, : 39 - 46
  • [6] A Data-Center FPGA Acceleration Platform for Convolutional Neural Networks
    Yu, Xiaoyu
    Wang, Yuwei
    Miao, Jie
    Zhang, Heng
    Meng, Yu
    Zhang, Bo
    Min, Biao
    Chen, Dewei
    Gao, Jianlin
    Wu, Ephrem
    2019 29TH INTERNATIONAL CONFERENCE ON FIELD-PROGRAMMABLE LOGIC AND APPLICATIONS (FPL), 2019, : 151 - 158
  • [7] CLASSIFICATION OF HUMAN GAIT ACCELERATION DATA USING CONVOLUTIONAL NEURAL NETWORKS
    Kreuter, Daniel
    Takahashi, Hirotaka
    Omae, Yuto
    Akiduki, Takuma
    Zhang, Zhong
    INTERNATIONAL JOURNAL OF INNOVATIVE COMPUTING INFORMATION AND CONTROL, 2020, 16 (02): : 609 - 619
  • [8] Artificial neural networks training acceleration through network science strategies
    Lucia Cavallaro
    Ovidiu Bagdasar
    Pasquale De Meo
    Giacomo Fiumara
    Antonio Liotta
    Soft Computing, 2020, 24 : 17787 - 17795
  • [9] Artificial Neural Networks Training Acceleration Through Network Science Strategies
    Cavallaro, Lucia
    Bagdasar, Ovidiu
    De Meo, Pasquale
    Fiumara, Giacomo
    Liotta, Antonio
    NUMERICAL COMPUTATIONS: THEORY AND ALGORITHMS, PT II, 2020, 11974 : 330 - 336
  • [10] Artificial neural networks training acceleration through network science strategies
    Cavallaro, Lucia
    Bagdasar, Ovidiu
    De Meo, Pasquale
    Fiumara, Giacomo
    Liotta, Antonio
    SOFT COMPUTING, 2020, 24 (23) : 17787 - 17795