Large Vocabulary SOUL Neural Network Language Models

被引:0
|
作者
Le, Hai-Son [1 ,2 ]
Oparin, Ilya [2 ]
Messaoudi, Abdel [2 ]
Allauzen, Alexandre [1 ,2 ]
Gauvain, Jean-Luc [2 ]
Yvon, Francois [1 ,2 ]
机构
[1] Univ Paris 11, BP 133, F-91403 Orsay, France
[2] Spoken Language Proc Grp, CNRS, LIMSI, F-91403 Orsay, France
关键词
Neural Network Language Model; Automatic Speech Recognition; Speech-To-Text;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper presents continuation of research on Structured OUt-put Layer Neural Network language models (SOUL NNLM) for automatic speech recognition. As SOUL NNLMs allow estimating probabilities for all in-vocabulary words and not only for those pertaining to a limited shortlist, we investigate. its performance on a large-vocabulary task. Significant improvements both in perplexity and word error rate over conventional shortlist-based NNLMs are shown on a challenging Arabic GALE task characterized by a recognition vocabulary of about 300k entries. A new training scheme is proposed for SOUL NNLMs that is based on separate training of the out-of-shortlist part of the output layer. It enables using more data at each iteration of a neural network without any considerable slow-down in training and brings additional improvements in speech recognition performance.
引用
收藏
页码:1480 / +
页数:2
相关论文
共 50 条
  • [1] Strategies for Training Large Vocabulary Neural Language Models
    Chen, Wenlin
    Grangier, David
    Auli, Michael
    PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, 2016, : 1975 - 1985
  • [2] Large Scale Hierarchical Neural Network Language Models
    Kuo, Hong-Kwang Jeff
    Arisoy, Ebru
    Emami, Ahmad
    Vozila, Paul
    13TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2012 (INTERSPEECH 2012), VOLS 1-3, 2012, : 1670 - 1673
  • [3] Unsupervised and Efficient Vocabulary Expansion for Recurrent Neural Network Language Models in ASR
    Khassanov, Yerbolat
    Chng, Eng Siong
    19TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2018), VOLS 1-6: SPEECH RESEARCH FOR EMERGING MARKETS IN MULTILINGUAL SOCIETIES, 2018, : 3343 - 3347
  • [4] A Factorized Recurrent Neural Network based architecture for medium to large vocabulary Language Modelling
    Iyer, Anantharaman Palacode Narayana
    2016 IEEE TENTH INTERNATIONAL CONFERENCE ON SEMANTIC COMPUTING (ICSC), 2016, : 292 - 299
  • [5] LEVERAGING LARGE LANGUAGE MODELS WITH VOCABULARY SHARING FOR SIGN LANGUAGE TRANSLATION
    Lee, Huije
    Kim, Jung-Ho
    Hwang, Eui Jun
    Kim, Jaewoo
    Park, Jong C.
    2023 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING WORKSHOPS, ICASSPW, 2023,
  • [6] PARAPHRASTIC LANGUAGE MODELS AND COMBINATION WITH NEURAL NETWORK LANGUAGE MODELS
    Liu, X.
    Gales, M. J. F.
    Woodland, P. C.
    2013 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2013, : 8421 - 8425
  • [7] Large vocabulary speech recognition with multispan statistical language models
    Bellegarda, JR
    IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, 2000, 8 (01): : 76 - 84
  • [8] The impact of chatbots based on large language models on second language vocabulary acquisition
    Zhang, Zhihui
    Huang, Xiaomeng
    HELIYON, 2024, 10 (03)
  • [9] Establishing vocabulary tests as a benchmark for evaluating large language models
    Martinez, Gonzalo
    Conde, Javier
    Merino-Gomez, Elena
    Bermudez-Margaretto, Beatriz
    Hernandez, Jose Alberto
    Reviriego, Pedro
    Brysbaert, Marc
    PLOS ONE, 2024, 19 (12):
  • [10] GNNAVI: Navigating the Information Flow in Large Language Models by Graph Neural Network
    Yuan, Shuzhou
    Nie, Ercong
    Faerber, Michael
    Schmid, Helmut
    Schuetze, Hinrich
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 3987 - 4001