Simultaneous neural machine translation with a reinforced attention mechanism

被引:6
|
作者
Lee, YoHan [1 ]
Shin, JongHun [1 ]
Kim, YoungKil [1 ]
机构
[1] Elect & Telecommun Res Inst, Language Intelligence Res Sect, Daejeon, South Korea
关键词
attention mechanism; neural network; reinforcement learning; simultaneous machine translation;
D O I
10.4218/etrij.2020-0358
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
To translate in real time, a simultaneous translation system should determine when to stop reading source tokens and generate target tokens corresponding to a partial source sentence read up to that point. However, conventional attention-based neural machine translation (NMT) models cannot produce translations with adequate latency in online scenarios because they wait until a source sentence is completed to compute alignment between the source and target tokens. To address this issue, we propose a reinforced learning (RL)-based attention mechanism, the reinforced attention mechanism, which allows a neural translation model to jointly train the stopping criterion and a partial translation model. The proposed attention mechanism comprises two modules, one to ensure translation quality and the other to address latency. Different from previous RL-based simultaneous translation systems, which learn the stopping criterion from a fixed NMT model, the modules can be trained jointly with a novel reward function. In our experiments, the proposed model has better translation quality and comparable latency compared to previous models.
引用
收藏
页码:775 / 786
页数:12
相关论文
共 50 条
  • [1] Fine-grained attention mechanism for neural machine translation
    Choi, Heeyoul
    Cho, Kyunghyun
    Bengio, Yoshua
    NEUROCOMPUTING, 2018, 284 : 171 - 176
  • [2] Attending From Foresight: A Novel Attention Mechanism for Neural Machine Translation
    Li, Xintong
    Liu, Lemao
    Tu, Zhaopeng
    Li, Guanlin
    Shi, Shuming
    Meng, Max Q. -H.
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2021, 29 : 2606 - 2616
  • [3] Recurrent Attention for Neural Machine Translation
    Zeng, Jiali
    Wu, Shuangzhi
    Yin, Yongjing
    Jiang, Yufan
    Li, Mu
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 3216 - 3225
  • [4] Neural Machine Translation with Deep Attention
    Zhang, Biao
    Xiong, Deyi
    Su, Jinsong
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2020, 42 (01) : 154 - 163
  • [5] Attention-via-Attention Neural Machine Translation
    Zhao, Shenjian
    Zhang, Zhihua
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 563 - 570
  • [6] Multilingual Simultaneous Neural Machine Translation
    Arthur, Philip
    Ryu, Dongwon K.
    Haffari, Gholamreza
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 4758 - 4766
  • [7] Neural machine translation for Indian language pair using hybrid attention mechanism
    Basab Nath
    Sunita Sarkar
    Surajeet Das
    Somnath Mukhopadhyay
    Innovations in Systems and Software Engineering, 2024, 20 : 175 - 183
  • [8] Neural machine translation for Indian language pair using hybrid attention mechanism
    Nath, Basab
    Sarkar, Sunita
    Das, Surajeet
    Mukhopadhyay, Somnath
    INNOVATIONS IN SYSTEMS AND SOFTWARE ENGINEERING, 2024, 20 (02) : 175 - 183
  • [9] Sparse and Constrained Attention for Neural Machine Translation
    Malaviya, Chaitanya
    Ferreira, Pedro
    Martins, Andre F. T.
    PROCEEDINGS OF THE 56TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 2, 2018, : 370 - 376
  • [10] Bilingual attention based neural machine translation
    Kang, Liyan
    He, Shaojie
    Wang, Mingxuan
    Long, Fei
    Su, Jinsong
    APPLIED INTELLIGENCE, 2023, 53 (04) : 4302 - 4315