A Non-Autoregressive Neural Machine Translation Model With Iterative Length Update of Target Sentence

被引:0
|
作者
Lim, Yeon-Soo [1 ]
Park, Eun-Ju [1 ]
Song, Hyun-Je [2 ]
Park, Seong-Bae [1 ]
机构
[1] Kyung Hee Univ, Dept Comp Sci & Engn, Yonin Si 17104, Gyeonggi Do, South Korea
[2] Jeonbuk Natl Univ, Dept Informat & Engn, Jeonju Si 54896, Jeollabuk Do, South Korea
基金
新加坡国家研究基金会;
关键词
Decoding; Iterative decoding; Generators; Adaptation models; Machine translation; Predictive models; Transformers; non-autoregressive decoder; sequence-to-sequence model; target length adaptation; transformer;
D O I
10.1109/ACCESS.2022.3169419
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The non-autoregressive decoders in neural machine translation are paid increasing attention due to their faster decoding than autoregressive decoders. However, their apparent problem is a low performance which is mainly originated from wrong prediction about the target sentence length. To attack this problem, this paper proposes a novel machine translation model with a new non-autoregressive decoder named Iterative and Length-Adjustive Non-Autoregressive Decoder (ILAND). This decoder adopts a masked language model to avoid generation of low-confident tokens and changes the length of a target sentence iteratively to an optimal length. To complete these goals, ILAND consists of three complementary sub-modules of a token masker, a length adjuster, and a token generator. The token masker and the token generator take charge of the masked language model, and the length adjuster optimizes the target sentence length. The sequence-to-sequence training of the translation model is also introduced. In this training, the length adjuster and the token generator are jointly trained since they share a similar structure. The effectiveness of the translation model is proven by showing empirically that the model outperforms other models with various non-autoregressive decoders. The thorough analysis suggests that the performance gain of the translation model comes from target sentence length adaptation and the joint learning. In addition, ILAND is also shown to be faster than other iterative non-autoregressive decoders while it is still robust against the multi-modality problem.
引用
收藏
页码:43341 / 43350
页数:10
相关论文
共 50 条
  • [21] Improving Non-Autoregressive Machine Translation Using Sentence-Level Semantic Agreement
    Wang, Shuheng
    Huang, Heyan
    Shi, Shumin
    APPLIED SCIENCES-BASEL, 2022, 12 (10):
  • [22] Guiding Non-Autoregressive Neural Machine Translation Decoding with Reordering Information
    Ran, Qiu
    Lin, Yankai
    Li, Peng
    Zhou, Jie
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 13727 - 13735
  • [23] Rephrasing the Reference for Non-autoregressive Machine Translation
    Shao, Chenze
    Zhang, Jinchao
    Zhou, Jie
    Feng, Yang
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 11, 2023, : 13538 - 13546
  • [24] Acyclic Transformer for Non-Autoregressive Machine Translation
    Huang, Fei
    Zhou, Hao
    Liu, Yang
    Li, Hang
    Huang, Minlie
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [25] Non-Autoregressive Machine Translation with Auxiliary Regularization
    Wang, Yiren
    Tian, Fei
    He, Di
    Qin, Tao
    Zhai, ChengXiang
    Liu, Tie-Yan
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 5377 - 5384
  • [26] Non-Autoregressive Machine Translation as Constrained HMM
    Li, Haoran
    Jie, Zhanming
    Lui, Wei
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 12361 - 12372
  • [27] Non-Autoregressive Machine Translation with Latent Alignments
    Saharia, Chitwan
    Chan, William
    Saxena, Saurabh
    Norouzi, Mohammad
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 1098 - 1108
  • [28] Non-Autoregressive Sentence Ordering
    Bin, Yi
    Shi, Wenhao
    Ji, Bin
    Zhang, Jipeng
    Ding, Yujuan
    Yang, Yang
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 4198 - 4214
  • [29] Enriching Non-Autoregressive Transformer with Syntactic and Semantic Structures for Neural Machine Translation
    Liu, Ye
    Wan, Yao
    Zhang, Jian-Guo
    Zhao, Wenting
    Yu, Philip S.
    16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 1235 - 1244
  • [30] Task-Level Curriculum Learning for Non-Autoregressive Neural Machine Translation
    Liu, Jinglin
    Ren, Yi
    Tan, Xu
    Zhang, Chen
    Qin, Tao
    Zhao, Zhou
    Liu, Tie-Yan
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 3861 - 3867