Composite decision by Bayesian inference in distant-talking speech recognition

被引:0
|
作者
Ji, Mikyong [1 ]
Kim, Sungtak [1 ]
Kim, Hoirin [1 ]
机构
[1] Informat & Commun Univ, SRT Lab, Taejon 305732, South Korea
来源
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper describes an integrated system to produce a composite recognition output on distant-talking speech when the recognition results from multiple microphone inputs are available. In many cases, the composite recognition result has lower error rate than any other individual output. In this work, the composite recognition result is obtained by applying Bayesian inference. The log likelihood score is assumed. to follow a Gaussian distribution, at least approximately. First, the distribution of the likelihood score is estimated in the development set. Then, the confidence interval for the likelihood score is used to remove unreliable microphone channels. Finally, the area under the distribution between the likelihood score of a hypothesis and that of the (N+1)(st) hypothesis is obtained for every channel and integrated for all channels by Bayesian inference. The proposed system shows considerable performance improvement compared with the result using an ordinary method by the summation of likelihoods as well as any of the recognition results of the channels.
引用
收藏
页码:463 / 470
页数:8
相关论文
共 50 条
  • [1] Robust distant-talking speech recognition
    Lin, Q
    Che, C
    Yuk, DS
    Jin, L
    deVries, B
    Pearson, J
    Flanagan, J
    1996 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, CONFERENCE PROCEEDINGS, VOLS 1-6, 1996, : 21 - 24
  • [2] Improved HMM separation for distant-talking speech recognition
    Takiguchi, T
    Nishimura, M
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2004, E87D (05): : 1127 - 1137
  • [3] ROBUSTNESS TO SPEAKER POSITION IN DISTANT-TALKING AUTOMATIC SPEECH RECOGNITION
    Gomez, Randy
    Nakamura, Keisuke
    Nakadai, Kazuhiro
    2013 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2013, : 7034 - 7038
  • [4] Environment-dependent denoising autoencoder for distant-talking speech recognition
    Ueda, Yuma
    Wang, Longbiao
    Kai, Atsuhiko
    Ren, Bo
    EURASIP JOURNAL ON ADVANCES IN SIGNAL PROCESSING, 2015,
  • [5] Environment-dependent denoising autoencoder for distant-talking speech recognition
    Yuma Ueda
    Longbiao Wang
    Atsuhiko Kai
    Bo Ren
    EURASIP Journal on Advances in Signal Processing, 2015
  • [6] Combination of bottleneck feature extraction and dereverberation for distant-talking speech recognition
    Bo Ren
    Longbiao Wang
    Liang Lu
    Yuma Ueda
    Atsuhiko Kai
    Multimedia Tools and Applications, 2016, 75 : 5093 - 5108
  • [7] Hidden Markov model training with contaminated speech material for distant-talking speech recognition
    Matassoni, M
    Omologo, M
    Giuliani, D
    Svaizer, P
    COMPUTER SPEECH AND LANGUAGE, 2002, 16 (02): : 205 - 223
  • [8] Combination of bottleneck feature extraction and dereverberation for distant-talking speech recognition
    Ren, Bo
    Wang, Longbiao
    Lu, Liang
    Ueda, Yuma
    Kai, Atsuhiko
    MULTIMEDIA TOOLS AND APPLICATIONS, 2016, 75 (09) : 5093 - 5108
  • [9] Denoising autoencoder and environment adaptation for distant-talking speech recognition with asynchronous speech recording
    Wang, Longbiao
    Ren, Bo
    Ueda, Yuma
    Kai, Atsuhiko
    Teraoka, Shunta
    Fukushima, Taku
    2014 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA), 2014,
  • [10] Experiments on distant-talking speech recognition in meeting room using extended MAM
    Pan, Y
    Waibel, A
    2002 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOLS I-IV, PROCEEDINGS, 2002, : 4165 - 4165