ZeUS: An Unified Training Framework for Constrained Neural Machine Translation

被引:0
|
作者
Yang, Murun [1 ]
机构
[1] Northeastern Univ, Sch Comp Sci & Engn, Shenyang 110819, Peoples R China
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Training; Data models; Synthetic data; Decoding; Neural machine translation; Terminology; Adaptation models; Neural networks; Translation; Constrained neural machine translation; large language model; synthetic data; unified training framework;
D O I
10.1109/ACCESS.2024.3454510
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Unlike general translation, constrained translation necessitates the proper use of predefined restrictions, such as specific terminologies and entities, during the translation process. However, current neural machine translation (NMT) models exhibit proficient performance solely in the domains of general translation or constrained translation. In this work, the author introduces the zero-shot unified constrained translation training framework, which adopts a novel approach of transforming constraints into textual explanations, thereby harmonizing the tasks of constrained translation with general translation. Furthermore, the author discovers the pivotal role of constructing synthetic data for domain-specific constrained translation in enhancing the model's performance on constrained translation tasks. To this end, the author utilizes large language models (LLMs) to generate domain-specific synthetic data for constrained translation. Experiments across four datasets and four translation directions, incorporating both general and constrained translations, demonstrate that models trained with the proposed framework and synthetic data achieve superior translation quality and constraint satisfaction rates, surpassing several baseline models in both general and contrained translation. Notably, ZeUS also exhibits significant advantages over multitask learning in constrained translation, with an average improvement of 7.25 percentage points in translation satisfaction rate (TSR) and 8.50 percentage points in translation completeness (TC).
引用
收藏
页码:124695 / 124704
页数:10
相关论文
共 50 条
  • [1] Restricted or Not: A General Training Framework for Neural Machine Translation
    Li, Zuchao
    Utiyama, Masao
    Sumita, Eiichiro
    Zhao, Hai
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022): STUDENT RESEARCH WORKSHOP, 2022, : 245 - 251
  • [2] Sparse and Constrained Attention for Neural Machine Translation
    Malaviya, Chaitanya
    Ferreira, Pedro
    Martins, Andre F. T.
    PROCEEDINGS OF THE 56TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 2, 2018, : 370 - 376
  • [3] Disambiguated Lexically Constrained Neural Machine Translation
    Zhang, Jinpeng
    Xia, Nini
    Wang, Ke
    Dong, Chuanqi
    Duan, Xiangyu
    Zhang, Yuqi
    Zhang, Min
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 10583 - 10596
  • [4] A Unified Neural Network for Quality Estimation of Machine Translation
    Li, Maoxi
    Xiang, Qingyu
    Chen, Zhiming
    Wang, Mingwen
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2018, E101D (09): : 2417 - 2421
  • [5] Lexically Constrained Neural Machine Translation with Levenshtein Transformer
    Susanto, Raymond Hendy
    Chollampatt, Shamil
    Tan, Liling
    58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), 2020, : 3536 - 3543
  • [6] Confidence Based Bidirectional Global Context Aware Training Framework for Neural Machine Translation
    Zhou, Chulun
    Meng, Fandong
    Zhou, Jie
    Zhang, Min
    Wang, Hongji
    Su, Jinsong
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 2878 - 2889
  • [7] A General Framework for Adaptation of Neural Machine Translation to Simultaneous Translation
    Chen, Yun
    Li, Liangyou
    Jiang, Xin
    Chen, Xiao
    Liu, Qun
    1ST CONFERENCE OF THE ASIA-PACIFIC CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 10TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (AACL-IJCNLP 2020), 2020, : 191 - 200
  • [8] Improving Neural Machine Translation by Bidirectional Training
    Ding, Liang
    Wu, Di
    Tao, Dacheng
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 3278 - 3284
  • [9] Discriminant training of neural networks for machine translation
    Quoc-Khanh Do
    Allauzen, Alexandre
    Yvon, Francois
    TRAITEMENT AUTOMATIQUE DES LANGUES, 2016, 57 (01): : 111 - 135
  • [10] Generative adversarial training for neural machine translation
    Yang, Zhen
    Chen, Wei
    Wang, Feng
    Xu, Bo
    NEUROCOMPUTING, 2018, 321 : 146 - 155