Probabilistic speech feature extraction with context-sensitive Bottleneck neural networks

被引:3
|
作者
Woellmer, Martin [1 ]
Schuller, Bjoern [1 ]
机构
[1] Tech Univ Munich, Inst Human Machine Commun, D-80333 Munich, Germany
关键词
Probabilistic feature extraction; Bottleneck networks; Long Short-Term Memory; Bidirectional speech processing; CONNECTIONIST FEATURE-EXTRACTION; BIDIRECTIONAL LSTM; NECK FEATURES;
D O I
10.1016/j.neucom.2012.06.064
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We introduce a novel context-sensitive feature extraction approach for spontaneous speech recognition. As bidirectional Long Short-Term Memory (BLSTM) networks are known to enable improved phoneme recognition accuracies by incorporating long-range contextual information into speech decoding, we integrate the BLSTM principle into a Tandem front-end for probabilistic feature extraction. Unlike the previously proposed approaches which exploit BLSTM modeling by generating a discrete phoneme prediction feature, our feature extractor merges continuous high-level probabilistic BLSTM features with low-level features. By combining BLSTM modeling and Bottleneck (BN) feature generation, we propose a novel front-end that allows us to produce context-sensitive probabilistic feature vectors of arbitrary size, independent of the network training targets. Evaluations on challenging spontaneous, conversational speech recognition tasks show that this concept prevails over recently published architectures for feature-level context modeling. (C) 2013 Elsevier B.V. All rights reserved.
引用
收藏
页码:113 / 120
页数:8
相关论文
共 50 条
  • [41] Context-sensitive data integration and prediction of biological networks
    Myers, Chad L.
    Troyanskaya, Olga G.
    BIOINFORMATICS, 2007, 23 (17) : 2322 - 2330
  • [42] Advanced Resource Provisioning in Context-Sensitive Converged Networks
    Castillo-Lema, J.
    Cruz, E.
    Neto, A.
    Cerqueira, E.
    2013 INTERNATIONAL CONFERENCE ON COMPUTING, NETWORKING AND COMMUNICATIONS (ICNC), 2013,
  • [43] Verification and Optimal Control of Context-Sensitive Probabilistic Boolean Networks Using Model Checking and Polynomial Optimization
    Kobayashi, Koichi
    Hiraishi, Kunihiko
    SCIENTIFIC WORLD JOURNAL, 2014,
  • [44] LEARNING RECURRENT NEURAL NETWORK LANGUAGE MODELS WITH CONTEXT-SENSITIVE LABEL SMOOTHING FOR AUTOMATIC SPEECH RECOGNITION
    Song, Minguang
    Zhao, Yunxin
    Wang, Shaojun
    Han, Mei
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 6159 - 6163
  • [45] Speech bottleneck feature extraction method based on overlapping group lasso sparse deep neural network
    Luo, Yuan
    Liu, Yu
    Zhang, Yi
    Yue, Congcong
    SPEECH COMMUNICATION, 2018, 99 : 56 - 61
  • [46] Probabilistic word classification based on a context-sensitive binary tree method
    Gao, J
    Chen, XX
    COMPUTER SPEECH AND LANGUAGE, 1997, 11 (04): : 307 - 320
  • [47] Efficient temporal probabilistic reasoning via context-sensitive model construction
    Ngo, L
    Haddawy, P
    Krieger, RA
    Helwig, J
    COMPUTERS IN BIOLOGY AND MEDICINE, 1997, 27 (05) : 453 - 476
  • [48] A context-sensitive technique for unsupervised change detection based on Hopfield-type neural networks
    Ghosh, Susmita
    Bruzzone, Lorenzo
    Patra, Swarnajyoti
    Bovolo, Francesca
    Ghosh, Ashish
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2007, 45 (03): : 778 - 789
  • [49] Transfer Learning with Bottleneck Feature Networks for Whispered Speech Recognition
    Lim, Boon Pang
    Wong, Faith
    Li, Yuyao
    Bay, Jia Wei
    17TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2016), VOLS 1-5: UNDERSTANDING SPEECH PROCESSING IN HUMANS AND MACHINES, 2016, : 1578 - 1582