TRAINING SAMPLE SELECTION FOR DEEP LEARNING OF DISTRIBUTED DATA

被引:0
|
作者
Jiang, Zheng [1 ]
Zhu, Xiaoqing [1 ]
Tan, Wai-tian [1 ]
Liston, Rob [1 ]
机构
[1] Cisco Syst, Chief Technol & Architecture Off, San Jose, CA 95134 USA
关键词
Deep neural networks; training sample selection; bandwidth-constrained learning;
D O I
暂无
中图分类号
TB8 [摄影技术];
学科分类号
0804 ;
摘要
The success of deep learning in the form of multi-layer neural networks depends critically on the volume and variety of training data. Its potential is greatly compromised when training data originate in a geographically distributed manner and are subject to bandwidth constraints. This paper presents a data sampling approach to deep learning, by carefully discriminating locally available training samples based on their relative importance. Towards this end, we propose two metrics for prioritizing candidate training samples as functions of their test trial outcome: correctness and confidence. Bandwidth-constrained simulations show significant performance gain of our proposed training sample selection schemes over convention uniform sampling: up to 15 x bandwidth reduction for the MNIST dataset and 25% reduction in learning time for the CIFAR-10 dataset.
引用
收藏
页码:2189 / 2193
页数:5
相关论文
共 50 条
  • [31] Research on the Deep Learning of the Small Sample Data based on Transfer Learning
    Zhao, Wei
    GREEN ENERGY AND SUSTAINABLE DEVELOPMENT I, 2017, 1864
  • [32] Impact of data set noise on distributed deep learning
    Qinghao G.
    Liguo S.
    Sunying H.
    Journal of China Universities of Posts and Telecommunications, 2020, 27 (02): : 37 - 45
  • [33] BigDL: A Distributed Deep Learning Framework for Big Data
    Dai, Jason
    Wang, Yiheng
    Qiu, Xin
    Ding, Ding
    Zhang, Yao
    Wang, Yanzhang
    Jia, Xianyan
    Zhang, Cherry
    Wan, Yan
    Li, Zhichao
    Wang, Jiao
    Huang, Shengsheng
    Wu, Zhongyuan
    Wang, Yang
    Yang, Yuhao
    She, Bowen
    Shi, Dongjie
    Lu, Qi
    Huang, Kai
    Song, Guoqiong
    PROCEEDINGS OF THE 2019 TENTH ACM SYMPOSIUM ON CLOUD COMPUTING (SOCC '19), 2019, : 50 - 60
  • [34] Distributed and deep vertical federated learning with big data
    Liu, Ji
    Zhou, Xuehai
    Mo, Lei
    Ji, Shilei
    Liao, Yuan
    Li, Zheng
    Gu, Qin
    Dou, Dejing
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2023, 35 (21):
  • [35] Impact of data set noise on distributed deep learning
    Guo Qinghao
    Shuai Liguo
    Hu Sunying
    TheJournalofChinaUniversitiesofPostsandTelecommunications, 2020, 27 (02) : 37 - 45
  • [36] Distributed Deep Learning for Remote Sensing Data Interpretation
    Haut, Juan M.
    Paoletti, Mercedes E.
    Moreno-Alvarez, Sergio
    Plaza, Javier
    Rico-Gallego, Juan-Antonio
    Plaza, Antonio
    PROCEEDINGS OF THE IEEE, 2021, 109 (08) : 1320 - 1349
  • [37] Distributed training strategies for a computer vision deep learning algorithm on a distributed GPU cluster
    Campos, Victor
    Sastre, Francesc
    Yagues, Maurici
    Bellver, Miriam
    Giro-i-Nieto, Xavier
    Torres, Jordi
    INTERNATIONAL CONFERENCE ON COMPUTATIONAL SCIENCE (ICCS 2017), 2017, 108 : 315 - 324
  • [38] A novel distributed deep learning training scheme based on distributed skip mesh list
    Suzuki, Masaya
    Mizutani, Kimihiro
    IEICE COMMUNICATIONS EXPRESS, 2021, 10 (08): : 463 - 468
  • [39] Evaluating the Informativity of a Training Sample for Image Classification by Deep Learning Methods
    Rusyn, B. P.
    Lutsyk, O. A.
    Kosarevych, R. Y.
    CYBERNETICS AND SYSTEMS ANALYSIS, 2021, 57 (06) : 853 - 863
  • [40] Dynamic Data Sample Selection and Scheduling in Edge Federated Learning
    Serhani, Mohamed Adel
    Abreha, Haftay Gebreslasie
    Tariq, Asadullah
    Hayajneh, Mohammad
    Xu, Yang
    Hayawi, Kadhim
    IEEE OPEN JOURNAL OF THE COMMUNICATIONS SOCIETY, 2023, 4 : 2133 - 2149