Lightweight and Elegant Data Reduction Strategies for Training Acceleration of Convolutional Neural Networks

被引:0
|
作者
Demidovskij, Alexander [1 ,2 ]
Tugaryov, Artyom [1 ]
Trutnev, Aleksei [1 ]
Kazyulina, Marina [1 ]
Salnikov, Igor [1 ]
Pavlov, Stanislav [1 ]
机构
[1] Huawei Russian Res Inst, NN AI Team, Ul Maksima Gorkogo 117, Nizhnii Novgorod 603006, Russia
[2] Natl Res Univ, Higher Sch Econ, Dept Informat Math & Comp Sci, Ul Bolshaya Pecherskaya 25-12, Nizhnii Novgorod 603155, Russia
关键词
deep learning training; training acceleration; convolutional neural networks; sample importance; dataset reduction;
D O I
10.3390/math11143120
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Due to industrial demands to handle increasing amounts of training data, lower the cost of computing one model at a time, and lessen the ecological effects of intensive computing resource consumption, the job of speeding the training of deep neural networks becomes exceedingly challenging. Adaptive Online Importance Sampling and IDS are two brand-new methods for accelerating training that are presented in this research. On the one hand, Adaptive Online Importance Sampling accelerates neural network training by lowering the number of forward and backward steps depending on how poorly a model can identify a given data sample. On the other hand, Intellectual Data Selection accelerates training by removing semantic redundancies from the training dataset and subsequently lowering the number of training steps. The study reports average 1.9x training acceleration for ResNet50, ResNet18, MobileNet v2 and YOLO v5 on a variety of datasets: CIFAR-100, CIFAR-10, ImageNet 2012 and MS COCO 2017, where training data are reduced by up to five times. Application of Adaptive Online Importance Sampling to ResNet50 training on ImageNet 2012 results in 2.37 times quicker convergence to 71.7% top-1 accuracy, which is within 5% of the baseline. Total training time for the same number of epochs as the baseline is reduced by 1.82 times, with an accuracy drop of 2.45 p.p. The amount of time required to apply Intellectual Data Selection to ResNet50 training on ImageNet 2012 is decreased by 1.27 times with a corresponding decline in accuracy of 1.12 p.p. Applying both methods to ResNet50 training on ImageNet 2012 results in 2.31 speedup with an accuracy drop of 3.5 p.p.
引用
收藏
页数:25
相关论文
共 50 条
  • [31] A Survey on Efficient Convolutional Neural Networks and Hardware Acceleration
    Ghimire, Deepak
    Kil, Dayoung
    Kim, Seong-heum
    ELECTRONICS, 2022, 11 (06)
  • [32] A Fourier domain acceleration framework for convolutional neural networks
    Lin, Jinhua
    Ma, Lin
    Yao, Yu
    NEUROCOMPUTING, 2019, 364 : 254 - 268
  • [33] Acceleration and implementation of convolutional neural networks based on FPGA
    Zhao, Sijie
    Gao, Shangshang
    Wang, Rugang
    Wang, Yuanyuan
    Zhou, Feng
    Guo, Naihong
    DIGITAL SIGNAL PROCESSING, 2023, 141
  • [34] Classification of seed corn ears based on custom lightweight convolutional neural network and improved training strategies
    Ma, Xiang
    Li, Yonglei
    Wan, Lipengcheng
    Xu, Zexin
    Song, Jiannong
    Huang, Jinqiu
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 120
  • [35] Training Data Reduction using Support Vectors for Neural Networks
    Tanio, Toranosuke
    Takeda, Kouya
    Yu, Jeahoon
    Hashimoto, Masanori
    2019 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2019, : 1405 - 1410
  • [36] Text Detection in Natural Images with Convolutional Neural Networks and Synthetic Training Data
    Grond, Marco
    Brink, Willie
    Herbst, Ben
    2016 PATTERN RECOGNITION ASSOCIATION OF SOUTH AFRICA AND ROBOTICS AND MECHATRONICS INTERNATIONAL CONFERENCE (PRASA-ROBMECH), 2016,
  • [37] Handwritten hiragana classifier with minimal training data utilizing Convolutional Neural Networks
    Chouza, Adrian
    Hernandez, Raul
    Jimenez, Jose
    Orozco-Rosas, Ulises
    Picos, Kenia
    OPTICS AND PHOTONICS FOR INFORMATION PROCESSING XIII, 2019, 11136
  • [38] JOINT TRAINING OF CONVOLUTIONAL AND NON-CONVOLUTIONAL NEURAL NETWORKS
    Soltau, Hagen
    Saon, George
    Sainath, Tara N.
    2014 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2014,
  • [39] Training Convolutional Neural Networks with Synthesized Data for Object Recognition in Industrial Manufacturing
    Li, Jason
    Gotvall, Per-Lage
    Provost, Julien
    Akesson, Knut
    2019 24TH IEEE INTERNATIONAL CONFERENCE ON EMERGING TECHNOLOGIES AND FACTORY AUTOMATION (ETFA), 2019, : 1544 - 1547
  • [40] Dealing with Lack of Training Data for Convolutional Neural Networks: The Case of Digital Pathology
    Ponzio, Francesco
    Urgese, Gianvito
    Ficarra, Elisa
    Di Cataldo, Santa
    ELECTRONICS, 2019, 8 (03)