Distant bigram language modelling using maximum entropy

被引:0
|
作者
Simons, M
Ney, H
Martin, SC
机构
关键词
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
In this paper, we apply tile maximum entropy approach to so-called distant bigram language modelling. In addition to the usual unigram and bigram dependencies, we use distant bigram dependencies, where tile immediate predecessor word of the word position under consideration is skipped. The contributions of this paper are: (1) We analyze the computational complexity of the resulting training algorithm, i.e. the generalized iterative scaling (GIS) algorithm, and studs the details of its implementation. (2) We describe a method for handling unseen events in the maximum entropy approach; this is achieved by discounting the frequencies of observed events. (3) We study the effect of this discounting operation on the convergence of the GIS algorithm. (4) We give experimental perplexity results for a corpus from the WSJ task. By using the maximum entropy approach and the distant bigram dependencies, we are able to reduce the perplexity from 205.4 for our best conventional bigram model to 169.5.
引用
收藏
页码:787 / 790
页数:4
相关论文
共 50 条
  • [31] A Maximum Entropy Modelling of the Rain Drop Size Distribution
    Checa, Ramiro
    Tapiador, Francisco J.
    ENTROPY, 2011, 13 (02) : 293 - 315
  • [32] Incorporating linguistic structure into maximum entropy language models
    GaoLin Fang
    Wen Gao
    ZhaoQi Wang
    Journal of Computer Science and Technology, 2003, 18 : 131 - 136
  • [33] Improved maximum entropy language model and its application
    Li, Juanzi
    Huang, Changning
    Ruan Jian Xue Bao/Journal of Software, 1999, 10 (03): : 257 - 263
  • [34] Latent maximum entropy principle for statistical language modeling
    Wang, SJ
    Rosenfeld, R
    Zhao, YX
    ASRU 2001: IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING, CONFERENCE PROCEEDINGS, 2001, : 182 - 185
  • [35] Learning to parse natural language with maximum entropy models
    Ratnaparkhi, A
    MACHINE LEARNING, 1999, 34 (1-3) : 151 - 175
  • [36] Digging Language Model - Maximum Entropy Phrase Extraction
    Kanis, Jakub
    TEXT, SPEECH, AND DIALOGUE, 2016, 9924 : 46 - 53
  • [37] Learning to Parse Natural Language with Maximum Entropy Models
    Adwait Ratnaparkhi
    Machine Learning, 1999, 34 : 151 - 175
  • [38] Incorporating linguistic structure into maximum entropy language models
    Fang, GL
    Gao, W
    Wang, ZQ
    JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY, 2003, 18 (01) : 131 - 136
  • [39] Maximum Entropy Named Entity Recognition for Czech Language
    Konkol, Michal
    Konopik, Miloslav
    TEXT, SPEECH AND DIALOGUE, TSD 2011, 2011, 6836 : 203 - 210
  • [40] Modulus maximum image energy using maximum entropy
    Heric, D
    Zazula, D
    PROCEEDINGS ELMAR-2005, 2005, : 57 - 60