Moment matching training for neural machine translation: An empirical study

被引:0
|
作者
Nguyen, Long H. B. [1 ,2 ]
Pham, Nghi T. [1 ,2 ]
Duc, Le D. C. [1 ,2 ]
Cong Duy Vu Hoang [3 ]
Dien Dinh [1 ,2 ]
机构
[1] Univ Sci Ho Chi Minh City, Fac Informat Technol, Ho Chi Minh City, Vietnam
[2] Vietnam Natl Univ, Ho Chi Minh City, Vietnam
[3] Oracle Corp, Melbourne, Vic, Australia
关键词
Neural machine translation; moment matching; objective function;
D O I
10.3233/JIFS-213240
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In recent years, Neural Machine Translation (NMT), which harnesses the power of neural networks, has achieved astonishing achievements. Despite its promise, NMT models can still not model prior external knowledge. Recent investigations have necessitated the adaptation of past expertise to both training and inference methods, resulting in translation inference issues. This paper proposes an extension of the moment matching framework that incorporates advanced prior knowledge without interfering with the inference process by using a matching mechanism between the model and empirical distributions. Our tests show that the suggested expansion outperforms the baseline and effectively over various language combinations.
引用
收藏
页码:2633 / 2645
页数:13
相关论文
共 50 条
  • [31] An Efficient Method for Generating Synthetic Data for Low-Resource Machine Translation An empirical study of Chinese, Japanese to Vietnamese Neural Machine Translation
    Thi-Vinh Ngo
    Phuong-Thai Nguyen
    Van Vinh Nguyen
    Thanh-Le Ha
    Le-Minh Nguyen
    APPLIED ARTIFICIAL INTELLIGENCE, 2022, 36 (01)
  • [32] Training with Adversaries to Improve Faithfulness of Attention in Neural Machine Translation
    Moradi, Pooya
    Kambhatla, Nishant
    Sarkar, Anoop
    AACL-IJCNLP 2020: THE 1ST CONFERENCE OF THE ASIA-PACIFIC CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 10TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING: PROCEEDINGS OF THE STUDENT RESEARCH WORKSHOP, 2020, : 86 - 93
  • [33] On the Copying Behaviors of Pre-Training for Neural Machine Translation
    Liu, Xuebo
    Wang, Longyue
    Wong, Derek F.
    Ding, Liang
    Chao, Lidia S.
    Shi, Shuming
    Tu, Zhaopeng
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 4265 - 4275
  • [34] Effectively training neural machine translation models with monolingual data
    Yang, Zhen
    Chen, Wei
    Wang, Feng
    Xu, Bo
    NEUROCOMPUTING, 2019, 333 : 240 - 247
  • [35] Token-level Adaptive Training for Neural Machine Translation
    Gu, Shuhao
    Zhang, Jinchao
    Meng, Fandong
    Feng, Yang
    Xie, Wanying
    Zhou, Jie
    Yu, Dong
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 1035 - 1046
  • [36] Curriculum pre-training for stylized neural machine translation
    Zou, Aixiao
    Wu, Xuanxuan
    Li, Xinjie
    Zhang, Ting
    Cui, Fuwei
    Xu, Jinan
    APPLIED INTELLIGENCE, 2024, 54 (17-18) : 7958 - 7968
  • [37] Bridging the Gap between Training and Inference for Neural Machine Translation
    Zhang, Wen
    Feng, Yang
    Liu, Qun
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 4790 - 4794
  • [38] Training with Additional Semantic Constraints for Enhancing Neural Machine Translation
    Ji, Yatu
    Hou, Hongxu
    Chen, Junjie
    Wu, Nier
    PRICAI 2019: TRENDS IN ARTIFICIAL INTELLIGENCE, PT I, 2019, 11670 : 300 - 313
  • [39] Synthetic Pre-Training Tasks for Neural Machine Translation
    He, Zexue
    Blackwood, Graeme
    Panda, Rameswar
    McAuley, Julian
    Feris, Rogerio
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 8080 - 8098
  • [40] Dynamic Sentence Sampling for Efficient Training of Neural Machine Translation
    Wang, Rui
    Utiyama, Masao
    Sumita, Eiichiro
    PROCEEDINGS OF THE 56TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 2, 2018, : 298 - 304