Converting Continuous-Space Language Models into N-gram Language Models with Efficient Bilingual Pruning for Statistical Machine Translation

被引:4
|
作者
Wang, Rui [1 ]
Utiyama, Masao [2 ]
Goto, Isao [3 ,4 ]
Sumita, Eiichiro [2 ]
Zhao, Hai [1 ,5 ]
Lu, Bao-Liang [1 ,5 ]
机构
[1] Shanghai Jiao Tong Univ, Ctr Brain Like Comp & Machine Intelligence, Dept Comp Sci & Engn, 800 Dongchuan Rd, Shanghai 200240, Peoples R China
[2] Natl Inst Informat & Commun Technol, Multilingual Translat Lab, 3-5 Hikaridai, Kyoto 6190289, Japan
[3] NHK Japan Broadcasting Corp, Sci & Technol Res Labs, Setagaya Ku, 1-10-11 Kinuta, Tokyo 1578510, Japan
[4] Natl Inst Informat & Commun Technol, Kyoto 6190289, Japan
[5] Shanghai Jiao Tong Univ, Key Lab Shanghai Educ Commiss Intelligent Interac, 800 Dongchuan Rd, Shanghai 200240, Peoples R China
基金
中国国家自然科学基金;
关键词
Machine translation; continuous-space language model; neural network language model; language model pruning;
D O I
10.1145/2843942
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The Language Model (LM) is an essential component of Statistical Machine Translation (SMT). In this article, we focus on developing efficient methods for LM construction. Our main contribution is that we propose a Natural N-grams based Converting (NNGC) method for transforming a Continuous-Space Language Model (CSLM) to a Back-off N-gram Language Model (BNLM). Furthermore, a Bilingual LM Pruning (BLMP) approach is developed for enhancing LMs in SMT decoding and speeding up CSLM converting. The proposed pruning and converting methods can convert a large LM efficiently by working jointly. That is, a LM can be effectively pruned before it is converted from CSLM without sacrificing performance, and further improved if an additional corpus contains out-of-domain information. For different SMT tasks, our experimental results indicate that the proposed NNGC and BLMP methods outperform the existing counterpart approaches significantly in BLEU and computational cost.
引用
收藏
页数:26
相关论文
共 50 条
  • [31] Efficient Estimation of Maximum Entropy Language Models with N-gram features: an SRILM extension
    Alumaee, Tanel
    Kurimo, Mikko
    11TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2010 (INTERSPEECH 2010), VOLS 3 AND 4, 2010, : 1820 - +
  • [32] An Investigation on Statistical Machine Translation with Neural Language Models
    Zhao, Yinggong
    Huang, Shujian
    Chen, Huadong
    Chen, Jiajun
    CHINESE COMPUTATIONAL LINGUISTICS AND NATURAL LANGUAGE PROCESSING BASED ON NATURALLY ANNOTATED BIG DATA, CCL 2014, 2014, 8801 : 175 - 186
  • [33] PERFORMANCE ANALYSIS OF NEURAL NETWORKS IN COMBINATION WITH N-GRAM LANGUAGE MODELS
    Oparin, Ilya
    Sundermeyer, Martin
    Ney, Hermann
    Gauvain, Jean-Luc
    2012 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2012, : 5005 - 5008
  • [34] Investigation on LSTM Recurrent N-gram Language Models for Speech Recognition
    Tueske, Zoltan
    Schlueter, Ralf
    Ney, Hermann
    19TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2018), VOLS 1-6: SPEECH RESEARCH FOR EMERGING MARKETS IN MULTILINGUAL SOCIETIES, 2018, : 3358 - 3362
  • [35] Rescore in a Flash: Compact, Cache Efficient Hashing Data Structures for N-gram Language Models
    Strimel, Grant P.
    Rastrow, Ariya
    Tiwari, Gautam
    Pierard, Adrien
    Webb, Jon
    INTERSPEECH 2020, 2020, : 3386 - 3390
  • [36] Combining naive Bayes and n-gram language models for text classification
    Peng, FC
    Schuurmans, D
    ADVANCES IN INFORMATION RETRIEVAL, 2003, 2633 : 335 - 350
  • [37] Stepwise API usage assistance using n-gram language models
    Santos, Andre L.
    Prendi, Goncalo
    Sousa, Hugo
    Ribeiro, Ricardo
    JOURNAL OF SYSTEMS AND SOFTWARE, 2017, 131 : 461 - 474
  • [38] N-gram Language Models in JLASER Neural Network Speech Recognizer
    Konopik, Miloslav
    Habernal, Ivan
    Brychcin, Tomas
    2010 INTERNATIONAL CONFERENCE ON APPLIED ELECTRONICS, 2010, : 167 - 170
  • [39] Bayesian learning of n-gram statistical language modeling
    Bai, Shuanhu
    Li, Haizhou
    2006 IEEE International Conference on Acoustics, Speech and Signal Processing, Vols 1-13, 2006, : 1045 - 1048
  • [40] Efficient Machine Translation Decoding with Slow Language Models
    Emami, Ahmad
    16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5, 2015, : 2376 - 2379