Subspace Based Sequence Discriminative Training of LSTM Acoustic Models with Feed-Forward Layers

被引:0
|
作者
Samarakoon, Lahiru [1 ]
Mak, Brian [2 ]
Lam, Albert Y. S. [1 ]
机构
[1] Fano Labs, Hong Kong, Peoples R China
[2] Hong Kong Univ Sci & Technol, Hong Kong, Peoples R China
关键词
Long Short-Term memory (LSTM); Recurrent Neural Networks (RNNs); Sequence Discriminative Training; Acoustic Modeling; NETWORKS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
State-of-the-art automatic speech recognition (ASR) systems use sequence discriminative training for improved performance over frame-level cross-entropy (CE) criterion. Even though sequence discriminative training improves long short-term memory (LSTM) recurrent neural network (RNN) acoustic models (AMs), it is not clear whether these systems achieve the optimal performance due to overfitting. This paper investigates the effect of state-level minimum Bayes risk (sMBR) training on LSTM AMs and shows that the conventional way of performing sMBR by updating all LSTM parameters is not optimal. We investigate two methods to improve the performance of sequence discriminative training of LSTM AMs. First more feedforward (FF) layers are included between the last LSTM layer and the output layer so those additional FF layers may benefit more from sMBR training. Second, a subspace is estimated as an interpolation of rank-1 matrices when performing sMBR for the LSTM layers of the AM. Our methods are evaluated in benchmark AMI single distance microphone (SDM) task. We find that the proposed approaches provide 1.6% absolute improvement over a strong sMBR trained LSTM baseline.
引用
收藏
页码:136 / 140
页数:5
相关论文
共 50 条
  • [21] Analytical interpretation of feed-forward nets outputs after training
    Garrido, L
    Gomez, S
    INTERNATIONAL JOURNAL OF NEURAL SYSTEMS, 1996, 7 (01) : 19 - 27
  • [22] Feed-forward neural network training using sparse representation
    Yang, Jie
    Ma, Jun
    EXPERT SYSTEMS WITH APPLICATIONS, 2019, 116 : 255 - 264
  • [23] In situ training of feed-forward and recurrent convolutional memristor networks
    Wang, Zhongrui
    Li, Can
    Lin, Peng
    Rao, Mingyi
    Nie, Yongyang
    Song, Wenhao
    Qiu, Qinru
    Li, Yunning
    Yan, Peng
    Strachan, John Paul
    Ge, Ning
    McDonald, Nathan
    Wu, Qing
    Hu, Miao
    Wu, Huaqiang
    Williams, R. Stanley
    Xia, Qiangfei
    Yang, J. Joshua
    NATURE MACHINE INTELLIGENCE, 2019, 1 (09) : 434 - 442
  • [24] Differential Evolution Training Algorithm for Feed-Forward Neural Networks
    Jarmo Ilonen
    Joni-Kristian Kamarainen
    Jouni Lampinen
    Neural Processing Letters, 2003, 17 : 93 - 105
  • [25] Multiplication units in feed-forward neural networks and its training
    Li, DZ
    Hirasawa, K
    Hu, JL
    Murata, J
    ICONIP'02: PROCEEDINGS OF THE 9TH INTERNATIONAL CONFERENCE ON NEURAL INFORMATION PROCESSING: COMPUTATIONAL INTELLIGENCE FOR THE E-AGE, 2002, : 75 - 79
  • [26] Feed-Forward Network Training Using Optimal Input Gains
    Malalur, Sanjeev S.
    Manry, Michael, Sr.
    IJCNN: 2009 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1- 6, 2009, : 2326 - 2333
  • [27] Constructing and training feed-forward neural networks for pattern classification
    Jiang, XD
    Wah, AHKS
    PATTERN RECOGNITION, 2003, 36 (04) : 853 - 867
  • [28] In situ training of feed-forward and recurrent convolutional memristor networks
    Zhongrui Wang
    Can Li
    Peng Lin
    Mingyi Rao
    Yongyang Nie
    Wenhao Song
    Qinru Qiu
    Yunning Li
    Peng Yan
    John Paul Strachan
    Ning Ge
    Nathan McDonald
    Qing Wu
    Miao Hu
    Huaqiang Wu
    R. Stanley Williams
    Qiangfei Xia
    J. Joshua Yang
    Nature Machine Intelligence, 2019, 1 : 434 - 442
  • [29] Training of large-scale feed-forward neural networks
    Seiffert, Udo
    2006 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORK PROCEEDINGS, VOLS 1-10, 2006, : 5324 - 5329
  • [30] Evolutionary approach to training feed-forward and recurrent neural networks
    Riley, Jeff
    Ciesielski, Victor B.
    International Conference on Knowledge-Based Intelligent Electronic Systems, Proceedings, KES, 1998, 3 : 596 - 602