Integration of context-dependent durational knowledge into HMM-based speech recognition

被引:0
|
作者
Wang, X
tenBosch, LFM
Pols, LCW
机构
关键词
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
This paper presents research on integrating context-dependent durational knowledge into HMM-based speech recognition. The first part of the paper presents work on obtaining relations between the parameters of the context-free HMMs and their durational behaviour, in preparation for the context-dependent durational modelling presented in the second part. Duration integration is realised via rescoring in the post-processing step of our N-best monophone recogniser. We use the multi-speaker TIMIT database for our analyses.
引用
收藏
页码:1073 / 1076
页数:4
相关论文
共 50 条
  • [21] An improved training algorithm in HMM-based speech recognition
    Li, GJ
    Huong, TY
    ICSLP 96 - FOURTH INTERNATIONAL CONFERENCE ON SPOKEN LANGUAGE PROCESSING, PROCEEDINGS, VOLS 1-4, 1996, : 1057 - 1060
  • [22] Scalable architecture for word HMM-based speech recognition
    Yoshizawa, S
    Wada, N
    Hayasaka, N
    Miyanaga, Y
    2004 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS, VOL 3, PROCEEDINGS, 2004, : 417 - 420
  • [23] Context-dependent hybrid HME/HMM speech recognition using polyphone clustering decision trees
    Fritsch, J
    Finke, M
    Waibel, A
    1997 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOLS I - V: VOL I: PLENARY, EXPERT SUMMARIES, SPECIAL, AUDIO, UNDERWATER ACOUSTICS, VLSI; VOL II: SPEECH PROCESSING; VOL III: SPEECH PROCESSING, DIGITAL SIGNAL PROCESSING; VOL IV: MULTIDIMENSIONAL SIGNAL PROCESSING, NEURAL NETWORKS - VOL V: STATISTICAL SIGNAL AND ARRAY PROCESSING, APPLICATIONS, 1997, : 1759 - 1762
  • [24] Bayesian Context Clustering Using Cross Valid Prior Distribution for HMM-Based Speech Recognition
    Hashimoto, Kei
    Zen, Heiga
    Nankaku, Yoshihiko
    Lee, Akinobu
    Tokuda, Keiichi
    INTERSPEECH 2008: 9TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2008, VOLS 1-5, 2008, : 936 - 939
  • [25] Use of acoustic contextual information in HMM-based speech recognition
    Korea Advanced Inst of Science and, Technology, Taejon, Korea, Republic of
    IEEE Signal Process Lett, 5 (108-110):
  • [26] Feature pruning in likelihood evaluation of HMM-based speech recognition
    Li, X
    Bilmes, J
    ASRU'03: 2003 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING ASRU '03, 2003, : 303 - 308
  • [27] From Stochastic Speech Recognition to Understanding: An HMM-Based approach
    Boda, PP
    1997 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING, PROCEEDINGS, 1997, : 57 - 64
  • [28] HMM-Based Speaker Emotional Recognition Technology for Speech Signal
    Qin, Yuqiang
    Zhang, Xueying
    FRONTIERS OF MANUFACTURING SCIENCE AND MEASURING TECHNOLOGY, PTS 1-3, 2011, 230-232 : 261 - 265
  • [29] Maximum likelihood linear transformations for HMM-based speech recognition
    Cambridge Univ Engineering Dep, Cambridge, United Kingdom
    Comput Speech Lang, 2 (75-98):
  • [30] Maximum likelihood linear transformations for HMM-based speech recognition
    Gales, MJF
    COMPUTER SPEECH AND LANGUAGE, 1998, 12 (02): : 75 - 98