Restricted or Not: A General Training Framework for Neural Machine Translation

被引:0
|
作者
Li, Zuchao [1 ,2 ]
Utiyama, Masao [3 ]
Sumita, Eiichiro [3 ]
Zhao, Hai [1 ,2 ]
机构
[1] Shanghai Jiao Tong Univ, Dept Comp Sci & Engn, Shanghai, Peoples R China
[2] Shanghai Jiao Tong Univ, AI Inst, MoE Key Lab Artificial Intelligence, Shanghai, Peoples R China
[3] Natl Inst Informat & Commun Technol NICT, Kyoto, Japan
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Restricted machine translation incorporates human prior knowledge into translation. It restricts the flexibility of the translation to satisfy the demands of translation in specific scenarios. Existing work typically imposes constraints on beam search decoding. Although this can satisfy the requirements overall, it usually requires a larger beam size and far longer decoding time than unrestricted translation, which limits the concurrent processing ability of the translation model in deployment, and thus its practicality. In this paper, we propose a general training framework that allows a model to simultaneously support both unrestricted and restricted translation by adopting an additional auxiliary training process without constraining the decoding process. This maintains the benefits of restricted translation but greatly reduces the extra time overhead of constrained decoding, thus improving its practicality. The effectiveness of our proposed training framework is demonstrated by experiments on both original (WAT21 En <-> Ja) and simulated (WMT14 En -> De and En -> Fr) restricted translation benchmarks.
引用
收藏
页码:245 / 251
页数:7
相关论文
共 50 条
  • [41] Neural Machine Translation
    Birch, Alexandra
    NATURAL LANGUAGE ENGINEERING, 2021, 27 (03) : 377 - 378
  • [42] Neural Machine Translation
    Jooste, Wandri
    Haque, Rejwanul
    Way, Andy
    MACHINE TRANSLATION, 2021, 35 (02) : 289 - 299
  • [43] DEEP: DEnoising Entity Pre-training for Neural Machine Translation
    Hu, Junjie
    Hayashi, Hiroaki
    Cho, Kyunghyun
    Neubig, Graham
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 1753 - 1766
  • [44] A Novel Framework for Neural Machine Translation of Indian-English Languages
    Nagarhalli, Tatwadarshi P.
    Vaze, Vinod
    Rana, N. K.
    PROCEEDINGS OF THE 5TH INTERNATIONAL CONFERENCE ON INVENTIVE COMPUTATION TECHNOLOGIES (ICICT-2020), 2020, : 676 - 682
  • [45] Neural Machine Translation Advised by Statistical Machine Translation
    Wang, Xing
    Lu, Zhengdong
    Tu, Zhaopeng
    Li, Hang
    Xiong, Deyi
    Zhang, Min
    THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 3330 - 3336
  • [46] C-NMT: A Collaborative Inference Framework for Neural Machine Translation
    Chen, Yukai
    Chiaro, Roberta
    Macii, Enrico
    Poncino, Massimo
    Pagliari, Daniele Jahier
    2022 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS 22), 2022, : 1512 - 1516
  • [47] Training and Inference Methods for High-Coverage Neural Machine Translation
    Yang, Michael
    Liu, Yixin
    Mayuranath, Rahul
    NEURAL GENERATION AND TRANSLATION, 2020, : 119 - 128
  • [48] Neural Machine Translation as a Novel Approach to Machine Translation
    Benkova, Lucia
    Benko, Lubomir
    DIVAI 2020: 13TH INTERNATIONAL SCIENTIFIC CONFERENCE ON DISTANCE LEARNING IN APPLIED INFORMATICS, 2020, : 499 - 508
  • [49] Data Rejuvenation: Exploiting Inactive Training Examples for Neural Machine Translation
    Jiao, Wenxiang
    Wang, Xing
    He, Shilin
    King, Irwin
    Lyu, Michael R.
    Tu, Zhaopeng
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 2255 - 2266
  • [50] Bilingual Mutual Information Based Adaptive Training for Neural Machine Translation
    Xu, Yangyifan
    Liu, Yijin
    Meng, Fandong
    Zhang, Jiajun
    Xu, Jinan
    Zhou, Jie
    ACL-IJCNLP 2021: THE 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 2, 2021, : 511 - 516