Whole-Word Recognition from Articulatory Movements for Silent Speech Interfaces

被引:0
|
作者
Wang, Jun [1 ]
Samal, Ashok
Green, Jordan R. [1 ]
Rudzicz, Frank
机构
[1] Univ Nebraska, Dept Special Educ & Commun Disorders, Lincoln, NE 68588 USA
关键词
silent speech recognition; speech impairment; laryngectomy; support vector machine; KNOWLEDGE;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Articulation-based silent speech interfaces convert silently produced speech movements into audible words. These systems are still in their experimental stages, but have significant potential for facilitating oral communication in persons with laryngectomy or speech impairments. In this paper, we report the result of a novel, real-time algorithm that recognizes whole-words based on articulatory movements. This approach differs from prior work that has focused primarily on phoneme-level recognition based on articulatory features. On average, our algorithm missed 1.93 words in a sequence of twenty-five words with an average latency of 0.79 seconds for each word prediction using a data set of 5,500 isolated word samples collected from ten speakers. The results demonstrate the effectiveness of our approach and its potential for building a real-time articulation-based silent speech interface for health applications.
引用
收藏
页码:1326 / 1329
页数:4
相关论文
共 50 条
  • [31] Web Interface for estimating articulatory movements in speech production from acoustics and text
    Udupa, Sathvik
    Roy, Anwesha
    Singh, Abhayjeet
    Illa, Aravind
    Ghosh, Prasanta Kumar
    INTERSPEECH 2021, 2021, : 4868 - 4869
  • [32] Enhanced recognition of written words and enjoyment of reading in struggling beginner readers through whole-word multimedia software
    Karemaker, Arjette
    Pitchford, Nicola J.
    O'Malley, Claire
    COMPUTERS & EDUCATION, 2010, 54 (01) : 199 - 208
  • [33] Speech Synthesis from Articulatory Movements Recorded by Real-time MRI
    Otani, Yuto
    Sawada, Shun
    Ohmura, Hidefumi
    Katsurada, Kouichi
    INTERSPEECH 2023, 2023, : 127 - 131
  • [34] Multi-Task Learning of Speech Recognition and Speech Synthesis Parameters for Ultrasound-based Silent Speech Interfaces
    Toth, Laszlo
    Gosztolya, Gabor
    Grosz, Tamas
    Marko, Alexandra
    Csapo, Tamas Gabor
    19TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2018), VOLS 1-6: SPEECH RESEARCH FOR EMERGING MARKETS IN MULTILINGUAL SOCIETIES, 2018, : 3172 - 3176
  • [36] MODEL OF WORD RECOGNITION FROM FLUENT SPEECH
    COLE, RA
    JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 1977, 62 : S49 - S49
  • [37] ARTICULATORY FEATURES FROM DEEP NEURAL NETWORKS AND THEIR ROLE IN SPEECH RECOGNITION
    Mitra, Vikramjit
    Sivaraman, Ganesh
    Nam, Hosung
    Espy-Wilson, Carol
    Saltzman, Elliot
    2014 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2014,
  • [38] On Acquiring Speech Production Knowledge from Articulatory Measurements for Phoneme Recognition
    Neiberg, D.
    Ananthakrishnan, G.
    Blomberg, M.
    INTERSPEECH 2009: 10TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2009, VOLS 1-5, 2009, : 1383 - 1386
  • [39] The development of whole-word representations in compound word processing: Evidence from eye fixation patterns of elementary school children
    Haikio, Tuomo
    Bertram, Raymond
    Hyona, Jukka
    APPLIED PSYCHOLINGUISTICS, 2011, 32 (03) : 533 - 551
  • [40] Estimation of articulatory movements from speech acoustics using an HMM-based speech production model
    Hiroya, S
    Honda, M
    IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, 2004, 12 (02): : 175 - 185