Lightweight and Elegant Data Reduction Strategies for Training Acceleration of Convolutional Neural Networks

被引:0
|
作者
Demidovskij, Alexander [1 ,2 ]
Tugaryov, Artyom [1 ]
Trutnev, Aleksei [1 ]
Kazyulina, Marina [1 ]
Salnikov, Igor [1 ]
Pavlov, Stanislav [1 ]
机构
[1] Huawei Russian Res Inst, NN AI Team, Ul Maksima Gorkogo 117, Nizhnii Novgorod 603006, Russia
[2] Natl Res Univ, Higher Sch Econ, Dept Informat Math & Comp Sci, Ul Bolshaya Pecherskaya 25-12, Nizhnii Novgorod 603155, Russia
关键词
deep learning training; training acceleration; convolutional neural networks; sample importance; dataset reduction;
D O I
10.3390/math11143120
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Due to industrial demands to handle increasing amounts of training data, lower the cost of computing one model at a time, and lessen the ecological effects of intensive computing resource consumption, the job of speeding the training of deep neural networks becomes exceedingly challenging. Adaptive Online Importance Sampling and IDS are two brand-new methods for accelerating training that are presented in this research. On the one hand, Adaptive Online Importance Sampling accelerates neural network training by lowering the number of forward and backward steps depending on how poorly a model can identify a given data sample. On the other hand, Intellectual Data Selection accelerates training by removing semantic redundancies from the training dataset and subsequently lowering the number of training steps. The study reports average 1.9x training acceleration for ResNet50, ResNet18, MobileNet v2 and YOLO v5 on a variety of datasets: CIFAR-100, CIFAR-10, ImageNet 2012 and MS COCO 2017, where training data are reduced by up to five times. Application of Adaptive Online Importance Sampling to ResNet50 training on ImageNet 2012 results in 2.37 times quicker convergence to 71.7% top-1 accuracy, which is within 5% of the baseline. Total training time for the same number of epochs as the baseline is reduced by 1.82 times, with an accuracy drop of 2.45 p.p. The amount of time required to apply Intellectual Data Selection to ResNet50 training on ImageNet 2012 is decreased by 1.27 times with a corresponding decline in accuracy of 1.12 p.p. Applying both methods to ResNet50 training on ImageNet 2012 results in 2.31 speedup with an accuracy drop of 3.5 p.p.
引用
收藏
页数:25
相关论文
共 50 条
  • [41] The Impact of Input Image Data Size on The Training Speed of Convolutional Neural Networks
    Lyu, Xinzhou
    2021 3RD INTERNATIONAL CONFERENCE ON MACHINE LEARNING, BIG DATA AND BUSINESS INTELLIGENCE (MLBDBI 2021), 2021, : 654 - 657
  • [42] Scaling up the training of Convolutional Neural Networks
    Snir, Marc
    2019 IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM WORKSHOPS (IPDPSW), 2019, : 925 - 925
  • [43] Towards dropout training for convolutional neural networks
    Wu, Haibing
    Gu, Xiaodong
    NEURAL NETWORKS, 2015, 71 : 1 - 10
  • [44] Bacterial classification with convolutional neural networks based on different data reduction layers
    Abd-Alhalem, Samia M.
    Soliman, Naglaa F.
    Abd Elrahman, Salah Eldin S. E.
    Ismail, Nabil A.
    El-Rabaie, El-Sayed M.
    Abd El-Samie, Fathi E.
    NUCLEOSIDES NUCLEOTIDES & NUCLEIC ACIDS, 2020, 39 (04): : 493 - 503
  • [45] CONVOLUTIONAL NEURAL NETWORKS TRAINING FOR AUTONOMOUS ROBOTICS
    Lozhkin, Alexander
    Maiorov, Konstantin
    Bozek, Pavol
    MANAGEMENT SYSTEMS IN PRODUCTION ENGINEERING, 2021, 29 (01) : 75 - 79
  • [46] DualConv: Dual Convolutional Kernels for Lightweight Deep Neural Networks
    Zhong, Jiachen
    Chen, Junying
    Mian, Ajmal
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (11) : 9528 - 9535
  • [47] Cow Rump Identification Based on Lightweight Convolutional Neural Networks
    Hou, Handan
    Shi, Wei
    Guo, Jinyan
    Zhang, Zhe
    Shen, Weizheng
    Kou, Shengli
    INFORMATION, 2021, 12 (09)
  • [48] Research Progress on Designing Lightweight Deep Convolutional Neural Networks
    Zhou, Zhifei
    Li, Hua
    Feng, Yixiong
    Lu, Jianguang
    Qian, Songrong
    Li, Shaobo
    Computer Engineering and Applications, 60 (22): : 1 - 17
  • [49] Lightweight Convolutional Neural Networks for CSI Feedback in Massive MIMO
    Cao, Zheng
    Shih, Wan-Ting
    Guo, Jiajia
    Wen, Chao-Kai
    Jin, Shi
    IEEE COMMUNICATIONS LETTERS, 2021, 25 (08) : 2624 - 2628
  • [50] IoT Malware Classification Based on Lightweight Convolutional Neural Networks
    Yuan, Baoguo
    Wang, Junfeng
    Wu, Peng
    Qing, Xianguo
    IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (05) : 3770 - 3783