Distributed Asynchronous Optimization of Convolutional Neural Networks

被引:0
|
作者
Chan, William [1 ]
Lane, Ian [1 ,2 ]
机构
[1] Carnegie Mellon Univ, Elect & Comp Engn, Pittsburgh, PA 15213 USA
[2] Carnegie Mellon Univ, Language Technol Inst, Pittsburgh, PA 15213 USA
关键词
deep neural network; distributed optimization;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, deep Convolutional Neural Networks have been shown to outperform Deep Neural Networks for acoustic modelling, producing state-of-the-art accuracy in speech recognition tasks. Convolutional models provide increased model robustness through the usage of pooling invariance and weight sharing across spectrum and time. However, training convolutional models is a very computationally expensive optimization procedure, especially when combined with large training corpora. In this paper, we present a novel algorithm for scalable training of deep Convolutional Neural Networks across multiple GPUs. Our distributed asynchronous stochastic gradient descent algorithm incorporates sparse gradients, momentum and gradient decay to accelerate the training of these networks. Our approach is stable, neither requiring warm-starting or excessively large minibatches. Our proposed approach enables convolutional models to be efficiently trained across multiple GPUs, enabling a model to be scaled asynchronously across 5 GPU workers with 68% efficiency.
引用
收藏
页码:1073 / 1077
页数:5
相关论文
共 50 条
  • [41] Distributed Convolutional Neural Networks for Human Activity Recognition in Wearable Robotics
    Hughes, Dana
    Correll, Nikolaus
    DISTRIBUTED AUTONOMOUS ROBOTIC SYSTEMS, 2019, 6 : 619 - 631
  • [42] Automatic Hyperparameter Tuning in Deep Convolutional Neural Networks Using Asynchronous Reinforcement Learning
    Neary, Patrick L.
    2018 IEEE INTERNATIONAL CONFERENCE ON COGNITIVE COMPUTING (ICCC), 2018, : 73 - 77
  • [43] Applications and Comparisons of Optimization Algorithms Used in Convolutional Neural Networks
    Seyyarer, Ebubekir
    Uckan, Taner
    Hark, Cengiz
    Ayata, Faruk
    Inan, Mevlut
    Karci, Ali
    2019 INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND DATA PROCESSING (IDAP 2019), 2019,
  • [44] Simplification and Optimization of Visual Tracking Convolutional Neural Networks Parameters
    Qin, Zhiyong
    Yu, Lixin
    PROCEEDINGS OF THE 2016 4TH INTERNATIONAL CONFERENCE ON MACHINERY, MATERIALS AND INFORMATION TECHNOLOGY APPLICATIONS, 2016, 71 : 253 - 259
  • [45] Speeding up the Hyperparameter Optimization of Deep Convolutional Neural Networks
    Hinz, Tobias
    Navarro-Guerrero, Nicolas
    Magg, Sven
    Wermter, Stefan
    INTERNATIONAL JOURNAL OF COMPUTATIONAL INTELLIGENCE AND APPLICATIONS, 2018, 17 (02)
  • [46] Communication Optimization for Distributed Execution of Graph Neural Networks
    Kurt, Sureyya Emre
    Yan, Jinghua
    Sukumaran-Rajam, Aravind
    Pandey, Prashant
    Sadayappan, P.
    2023 IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM, IPDPS, 2023, : 512 - 523
  • [47] Asynchronous Optimization Methods for Efficient Training of Deep Neural Networks with Guarantees
    Kungurtsev, Vyacheslav
    Egan, Malcolm
    Chatterjee, Bapi
    Alistarh, Dan
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 8209 - 8216
  • [48] Convolutional Neural Networks Based Motion Data Optimization Networks for Leap Motion
    Zhang X.
    Xie W.
    Li S.
    Liu X.
    Jisuanji Fuzhu Sheji Yu Tuxingxue Xuebao/Journal of Computer-Aided Design and Computer Graphics, 2021, 33 (03): : 439 - 447
  • [49] Convolutional neural networks
    Alexander Derry
    Martin Krzywinski
    Naomi Altman
    Nature Methods, 2023, 20 : 1269 - 1270
  • [50] Convolutional neural networks
    Derry, Alexander
    Krzywinski, Martin
    Altman, Naomi
    NATURE METHODS, 2023, 20 (09) : 1269 - 1270