TRAINING SAMPLE SELECTION FOR DEEP LEARNING OF DISTRIBUTED DATA

被引:0
|
作者
Jiang, Zheng [1 ]
Zhu, Xiaoqing [1 ]
Tan, Wai-tian [1 ]
Liston, Rob [1 ]
机构
[1] Cisco Syst, Chief Technol & Architecture Off, San Jose, CA 95134 USA
关键词
Deep neural networks; training sample selection; bandwidth-constrained learning;
D O I
暂无
中图分类号
TB8 [摄影技术];
学科分类号
0804 ;
摘要
The success of deep learning in the form of multi-layer neural networks depends critically on the volume and variety of training data. Its potential is greatly compromised when training data originate in a geographically distributed manner and are subject to bandwidth constraints. This paper presents a data sampling approach to deep learning, by carefully discriminating locally available training samples based on their relative importance. Towards this end, we propose two metrics for prioritizing candidate training samples as functions of their test trial outcome: correctness and confidence. Bandwidth-constrained simulations show significant performance gain of our proposed training sample selection schemes over convention uniform sampling: up to 15 x bandwidth reduction for the MNIST dataset and 25% reduction in learning time for the CIFAR-10 dataset.
引用
收藏
页码:2189 / 2193
页数:5
相关论文
共 50 条
  • [1] Deep Learning for Interference Identification: Band, Training SNR, and Sample Selection
    Zhang, Xiwen
    Seyfi, Tolunay
    Ju, Shengtai
    Ramjee, Sharan
    El Gamal, Aly
    Eldar, Yonina C.
    2019 IEEE 20TH INTERNATIONAL WORKSHOP ON SIGNAL PROCESSING ADVANCES IN WIRELESS COMMUNICATIONS (SPAWC 2019), 2019,
  • [2] Training sample selection in learning control
    Cheng, J
    Xu, YS
    Chung, R
    IEEE ROBIO 2004: PROCEEDINGS OF THE IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS, 2004, : 368 - 373
  • [3] Exploring the Effects of Silent Data Corruption in Distributed Deep Learning Training
    Rojas, Elvis
    Perez, Diego
    Meneses, Esteban
    2022 IEEE 34TH INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE AND HIGH PERFORMANCE COMPUTING (SBAC-PAD 2022), 2022, : 21 - 30
  • [4] Incorporating Reinforcement Learning for Quality-aware Sample Selection in Deep Architecture Training
    Adhane, Gereziher
    Dehshibi, Mohammad Mahdi
    Masip, David
    2022 IEEE INTERNATIONAL CONFERENCE ON OMNI-LAYER INTELLIGENT SYSTEMS (IEEE COINS 2022), 2022, : 413 - 417
  • [5] Survey on Network of Distributed Deep Learning Training
    Zhu H.
    Yuan G.
    Yao C.
    Tan G.
    Wang Z.
    Hu Z.
    Zhang X.
    An X.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2021, 58 (01): : 98 - 115
  • [6] Selective Preemption of Distributed Deep Learning Training
    Go, Younghun
    Shin, Changyong
    Lee, Jeunghwan
    Yoo, Yeonho
    Yang, Gyeongsik
    Yoo, Chuck
    2023 IEEE 16TH INTERNATIONAL CONFERENCE ON CLOUD COMPUTING, CLOUD, 2023, : 175 - 177
  • [7] Two-stage selection of distributed data centers based on deep reinforcement learning
    Li, Qirui
    Peng, Zhiping
    Cui, Delong
    Lin, Jianpeng
    He, Jieguang
    Cluster Computing, 2022, 25 (04) : 2699 - 2714
  • [8] Two-stage selection of distributed data centers based on deep reinforcement learning
    Li, Qirui
    Peng, Zhiping
    Cui, Delong
    Lin, Jianpeng
    He, Jieguang
    CLUSTER COMPUTING-THE JOURNAL OF NETWORKS SOFTWARE TOOLS AND APPLICATIONS, 2022, 25 (04): : 2699 - 2714
  • [9] Two-stage selection of distributed data centers based on deep reinforcement learning
    Qirui Li
    Zhiping Peng
    Delong Cui
    Jianpeng Lin
    Jieguang He
    Cluster Computing, 2022, 25 : 2699 - 2714
  • [10] Analyzing he distributed training of deep-learning models via data locality
    Alonso-Monsalve, Saul
    Calderon, Alejandro
    Garcia-Carballeira, Felix
    Rivadeneira, Jose
    2021 29TH EUROMICRO INTERNATIONAL CONFERENCE ON PARALLEL, DISTRIBUTED AND NETWORK-BASED PROCESSING (PDP 2021), 2021, : 117 - 121