UNNORMALIZED EXPONENTIAL AND NEURAL NETWORK LANGUAGE MODELS

被引:0
|
作者
Sethy, Abhinav [1 ]
Chen, Stanley [1 ]
Arisoy, Ebru [1 ]
Ramabhadran, Bhuvana [1 ]
机构
[1] IBM Corp, TJ Watson Res Ctr, Yorktown Hts, NY 10598 USA
关键词
Model M; unnormalized models; neural network language models; fast lookup;
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Model M, an exponential class-based language model, and neural network language models (NNLM's) have outperformed word n-gram language models over a wide range of tasks. However, these gains come at the cost of vastly increased computation when calculating word probabilities. For both models, the bulk of this computation involves evaluating the softmax function over a large word or class vocabulary to ensure that probabilities sum to 1. In this paper, we study unnormalized variants of Model M and NNLM's, whereby the softmax function is simply omitted. Accordingly, model training must be modified to encourage scores to sum close to 1. In this paper, we demonstrate up to a factor of 35 faster n-gram lookups with unnormalized models over their normalized counterparts, while still yielding state-of-the-art performance in WER (10.2 on the English broadcast news rt04 set).
引用
收藏
页码:5416 / 5420
页数:5
相关论文
共 50 条
  • [21] Finding Fuzziness in Neural Network Models of Language Processing
    Misra, Kanishka
    Rayz, Julia Taylor
    EXPLAINABLE AI AND OTHER APPLICATIONS OF FUZZY TECHNIQUES, NAFIPS 2021, 2022, 258 : 278 - 290
  • [22] Neural Network Language Models for Translation with United Data
    Khalilov, Maxim
    Fonollosa, Jose A. R.
    Zamora-Martinez, R.
    Castro-Bleda, M. J.
    Espana-Boquera, S.
    20TH IEEE INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, VOL 2, PROCEEDINGS, 2008, : 445 - +
  • [23] Neural network models for language acquisition: A brief survey
    Poveda, Jordi
    Vellido, Alfredo
    INTELLIGENT DATA ENGINEERING AND AUTOMATED LEARNING - IDEAL 2006, PROCEEDINGS, 2006, 4224 : 1346 - 1357
  • [24] Machine Translation based on Neural Network Language Models
    Zamora-Martinez, Francisco
    Jose Castro-Bleda, Maria
    PROCESAMIENTO DEL LENGUAJE NATURAL, 2010, (45): : 221 - 228
  • [25] Large Scale Hierarchical Neural Network Language Models
    Kuo, Hong-Kwang Jeff
    Arisoy, Ebru
    Emami, Ahmad
    Vozila, Paul
    13TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2012 (INTERSPEECH 2012), VOLS 1-3, 2012, : 1670 - 1673
  • [26] OPTIMIZATION OF NEURAL NETWORK LANGUAGE MODELS FOR KEYWORD SEARCH
    Gandhe, Ankur
    Metze, Florian
    Waibel, Alex
    Lane, Ian
    2014 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2014,
  • [27] COMPARISON OF FEEDFORWARD AND RECURRENT NEURAL NETWORK LANGUAGE MODELS
    Sundermeyer, M.
    Oparin, I.
    Gauvain, J. -L.
    Freiberg, B.
    Schlueter, R.
    Ney, H.
    2013 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2013, : 8430 - 8434
  • [28] FUTURE WORD CONTEXTS IN NEURAL NETWORK LANGUAGE MODELS
    Chen, X.
    Liu, X.
    Ragni, A.
    Wang, Y.
    Gales, M. J. F.
    2017 IEEE AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING WORKSHOP (ASRU), 2017, : 97 - 103
  • [29] Global exponential stability of nonautonomous neural network models with unbounded delays
    Oliveira, Jose J.
    NEURAL NETWORKS, 2017, 96 : 71 - 79
  • [30] Impulsive effects on the global exponential stability of neural network models with supremums
    Stamova, Ivanka M.
    Stamov, Trayan
    Simeonova, Neli
    EUROPEAN JOURNAL OF CONTROL, 2014, 20 (04) : 199 - 206