Multi-softmax Deep Neural Network for Semi-supervised Training

被引:0
|
作者
Su, Hang [1 ,2 ]
Xu, Haihua [3 ]
机构
[1] Int Comp Sci Inst, Berkeley, CA 94704 USA
[2] Univ Calif Berkeley, Dept Elect Engn & Comp Sci, Berkeley, CA 94720 USA
[3] Nanyang Technol Univ, Singapore, Singapore
关键词
Semi-supervised training; Low resources; Deep Neural Networks;
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
In this paper we propose a Shared Hidden Layer Multi-softmax Deep Neural Network (SHL-MDNN) approach for semi-supervised training (SST). This approach aims to boost low-resource speech recognition where limited training data is available. Supervised data and unsupervised data share the same hidden layers but are fed into different softmax layers so that erroneous automatic speech recognition (ASR) transcriptions of the unsupervised data have less effect on shared hidden layers. Experimental results on Babel data indicate that this approach always outperform naive SST on DNN, and it can yield 1.3% word error rate (WER) reduction compared with supervised DNN hybrid system. In addition, if softmax layer is retrained with supervised data, it can lead up to another 0.8% WER reduction. Confidence based data selection is also studied in this setup. Experiments show that this method is not sensitive to ASR transcription errors.
引用
收藏
页码:3239 / 3243
页数:5
相关论文
共 50 条
  • [1] SEMI-SUPERVISED TRAINING OF DEEP NEURAL NETWORKS
    Vesely, Karel
    Hannemann, Mirko
    Burget, Lukas
    2013 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING (ASRU), 2013, : 267 - 272
  • [2] SEMI-SUPERVISED TRAINING STRATEGIES FOR DEEP NEURAL NETWORKS
    Gibson, Matthew
    Cook, Gary
    Zhan, Puming
    2017 IEEE AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING WORKSHOP (ASRU), 2017, : 77 - 83
  • [3] Semi-supervised Maximum Mutual Information Training of Deep Neural Network Acoustic Models
    Manohar, Vimal
    Povey, Daniel
    Khudanpur, Sanjeev
    16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5, 2015, : 2630 - 2634
  • [4] DEEP NEURAL NETWORK FEATURES AND SEMI-SUPERVISED TRAINING FOR LOW RESOURCE SPEECH RECOGNITION
    Thomas, Samuel
    Seltzer, Michael L.
    Church, Kenneth
    Hermansky, Hynek
    2013 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2013, : 6704 - 6708
  • [5] Decoupled Deep Neural Network for Semi-supervised Semantic Segmentation
    Hong, Seunghoon
    Noh, Hyeonwoo
    Han, Bohyung
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 28 (NIPS 2015), 2015, 28
  • [6] SemiH: DFT Hamiltonian neural network training with semi-supervised learning
    Cho, Yucheol
    Choi, Guenseok
    Ham, Gyeongdo
    Shin, Mincheol
    Kim, Daeshik
    MACHINE LEARNING-SCIENCE AND TECHNOLOGY, 2024, 5 (03):
  • [7] SEMI-SUPERVISED BOOTSTRAPPING APPROACH FOR NEURAL NETWORK FEATURE EXTRACTOR TRAINING
    Grezl, Frantisek
    Karafiat, Martin
    2013 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING (ASRU), 2013, : 470 - 475
  • [8] Semi-Supervised Adversarial Training of a Lightweight Neural Network for Visual Recognition
    Karaman, Kaan
    Akkaya, Ibrahim Batuhan
    COUNTERTERRORISM, CRIME FIGHTING, FORENSICS, AND SURVEILLANCE TECHNOLOGIES III, 2019, 11166
  • [9] A Neural Network for Semi-supervised Learning on Manifolds
    Genkin, Alexander
    Sengupta, Anirvan M.
    Chklovskii, Dmitri
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2019: THEORETICAL NEURAL COMPUTATION, PT I, 2019, 11727 : 375 - 386
  • [10] Labeling Malicious Communication Samples Based on Semi-Supervised Deep Neural Network
    Shao, Guolin
    Chen, Xingshu
    Zeng, Xuemei
    Wang, Lina
    CHINA COMMUNICATIONS, 2019, 16 (11) : 183 - 200