Data selection by sequence summarizing neural network in mismatch condition training

被引:2
|
作者
Zmolikova, Katerina [1 ,2 ]
Karafiat, Martin [1 ,2 ]
Vesely, Karel [1 ,2 ]
Delcroix, Marc [3 ]
Watanabe, Shinji [4 ]
Burget, Lukas [1 ,2 ]
Cernocky, Jan Honza [1 ,2 ]
机构
[1] Brno Univ Technol, Speech FIT, Brno, Czech Republic
[2] IT4I Ctr Excellence, Brno, Czech Republic
[3] NTT Corp, NTT Commun Sci Labs, Kyoto, Japan
[4] MERL, Cambridge, MA USA
关键词
Automatic speech recognition; Data augmentation; Data selection; Mismatch training condition; Sequence summarization;
D O I
10.21437/Interspeech.2016-741
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Data augmentation is a simple and efficient technique to improve the robustness of a speech recognizer when deployed in mismatched training-test conditions. Our paper proposes a new approach for selecting data with respect to similarity of acoustic conditions. The similarity is computed based on a sequence summarizing neural network which extracts vectors containing acoustic summary (e.g. noise and reverberation characteristics) of an utterance. Several configurations of this network and different methods of selecting data using these "summary-vectors" were explored. The results are reported on a mismatched condition using AMI training set with the proposed data selection and CHiME3 test set.
引用
收藏
页码:2354 / 2358
页数:5
相关论文
共 50 条
  • [21] A novel neural network training framework with data assimilation
    Chen, Chong
    Dou, Yixuan
    Chen, Jie
    Xue, Yaru
    JOURNAL OF SUPERCOMPUTING, 2022, 78 (17): : 19020 - 19045
  • [22] A novel neural network training framework with data assimilation
    Chong Chen
    Yixuan Dou
    Jie Chen
    Yaru Xue
    The Journal of Supercomputing, 2022, 78 : 19020 - 19045
  • [23] Measuring the Effects of Data Parallelism on Neural Network Training
    Shallue, Christopher J.
    Lee, Jaehoon
    Antognini, Joseph
    Sohl-Dickstein, Jascha
    Frostig, Roy
    Dahl, George E.
    JOURNAL OF MACHINE LEARNING RESEARCH, 2019, 20
  • [24] Accelerating Data Loading in Deep Neural Network Training
    Yang, Chih-Chieh
    Cong, Guojing
    2019 IEEE 26TH INTERNATIONAL CONFERENCE ON HIGH PERFORMANCE COMPUTING, DATA, AND ANALYTICS (HIPC), 2019, : 235 - 245
  • [25] Training data selection method for generalization by multilayer neural networks
    Hara, K
    Nakayama, K
    IEICE TRANSACTIONS ON FUNDAMENTALS OF ELECTRONICS COMMUNICATIONS AND COMPUTER SCIENCES, 1998, E81A (03) : 374 - 381
  • [26] MAXIMUM COVERAGE METHOD FOR FEATURE SUBSET SELECTION FOR NEURAL NETWORK TRAINING
    Boor, Stefan
    COMPUTING AND INFORMATICS, 2011, 30 (05) : 901 - 912
  • [27] Selection of optimal parameter of speed of training of neural network perceptron type
    Kulchin, YN
    Denisova, EV
    Denisov, IV
    Efimov, MB
    FUNDAMENTAL PROBLEMS OF OPTOELECTRONICS AND MICROELECTRONICS, 2003, 5129 : 162 - 167
  • [28] Samples Selection for Artificial Neural Network Training in Preliminary Structural Design
    童霏
    刘西拉
    Tsinghua Science and Technology, 2005, (02) : 233 - 239
  • [29] Adaptive key frames selection algorithms for summarizing video data
    Farag, WE
    Abdel-Wahab, H
    PROCEEDINGS OF THE 6TH JOINT CONFERENCE ON INFORMATION SCIENCES, 2002, : 1017 - 1020
  • [30] Issues in training set selection and refinement for classification by a feedforward neural network
    Foody, GM
    IGARSS '98 - 1998 INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM, PROCEEDINGS VOLS 1-5: SENSING AND MANAGING THE ENVIRONMENT, 1998, : 409 - 411