TRAINING SAMPLE SELECTION FOR DEEP LEARNING OF DISTRIBUTED DATA

被引:0
|
作者
Jiang, Zheng [1 ]
Zhu, Xiaoqing [1 ]
Tan, Wai-tian [1 ]
Liston, Rob [1 ]
机构
[1] Cisco Syst, Chief Technol & Architecture Off, San Jose, CA 95134 USA
关键词
Deep neural networks; training sample selection; bandwidth-constrained learning;
D O I
暂无
中图分类号
TB8 [摄影技术];
学科分类号
0804 ;
摘要
The success of deep learning in the form of multi-layer neural networks depends critically on the volume and variety of training data. Its potential is greatly compromised when training data originate in a geographically distributed manner and are subject to bandwidth constraints. This paper presents a data sampling approach to deep learning, by carefully discriminating locally available training samples based on their relative importance. Towards this end, we propose two metrics for prioritizing candidate training samples as functions of their test trial outcome: correctness and confidence. Bandwidth-constrained simulations show significant performance gain of our proposed training sample selection schemes over convention uniform sampling: up to 15 x bandwidth reduction for the MNIST dataset and 25% reduction in learning time for the CIFAR-10 dataset.
引用
收藏
页码:2189 / 2193
页数:5
相关论文
共 50 条
  • [41] Evaluating the Informativity of a Training Sample for Image Classification by Deep Learning Methods
    B. P. Rusyn
    O. A. Lutsyk
    R. Y. Kosarevych
    Cybernetics and Systems Analysis, 2021, 57 : 853 - 863
  • [42] Incorporating Negative Sample Training for Ship Detection Based on Deep Learning
    Gao, Lianru
    He, Yiqun
    Sun, Xu
    Jia, Xiuping
    Zhang, Bing
    SENSORS, 2019, 19 (03)
  • [43] LOW: Training deep neural networks by learning optimal sample weights
    Santiago, Carlos
    Barata, Catarina
    Sasdelli, Michele
    Carneiro, Gustavo
    Nascimento, Jacinto C.
    Pattern Recognition, 2021, 110
  • [44] Learning With Imbalanced Noisy Data by Preventing Bias in Sample Selection
    Liu, Huafeng
    Sheng, Mengmeng
    Sun, Zeren
    Yao, Yazhou
    Hua, Xian-Sheng
    Shen, Heng-Tao
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 7426 - 7437
  • [45] LOW: Training deep neural networks by learning optimal sample weights
    Santiago, Carlos
    Barata, Catarina
    Sasdelli, Michele
    Carneiro, Gustavo
    Nascimento, Jacinto C.
    PATTERN RECOGNITION, 2021, 110
  • [46] Learning from Imbalanced Data Using Methods of Sample Selection
    Chairi, Ikram
    Alaoui, Souad
    Lyhyaoui, Abdelouahid
    2012 INTERNATIONAL CONFERENCE ON MULTIMEDIA COMPUTING AND SYSTEMS (ICMCS), 2012, : 256 - 259
  • [47] Deep Learning for Proteomics Data for Feature Selection and Classification
    Iravani, Sahar
    Conrad, Tim O. F.
    MACHINE LEARNING AND KNOWLEDGE EXTRACTION, CD-MAKE 2019, 2019, 11713 : 301 - 316
  • [48] SelectNet: Learning to Sample from the Wild for Imbalanced Data Training
    Liu, Yunru
    Gao, Tingran
    Yang, Haizhao
    MATHEMATICAL AND SCIENTIFIC MACHINE LEARNING, VOL 107, 2020, 107 : 193 - 206
  • [49] Bounding Training Data Reconstruction in Private (Deep) Learning
    Guo, Chuan
    Karrer, Brian
    Chaudhuri, Kamalika
    van der Maaten, Laurens
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [50] Synthetic seismic data for training deep learning networks
    Merrifield, Tom P.
    Griffith, Donald P.
    Zamanian, S. Ahmad
    Gesbert, Stephane
    Sen, Satyakee
    Guzman, Jorge De La Torre
    Potter, R. David
    Kuehl, Henning
    INTERPRETATION-A JOURNAL OF SUBSURFACE CHARACTERIZATION, 2022, 10 (03): : SE31 - SE39