Neural Machine Translation with Key-Value Memory-Augmented Attention

被引:0
|
作者
Meng, Fandong [1 ]
Tu, Zhaopeng [1 ]
Cheng, Yong [1 ]
Wu, Haiyang [1 ]
Zhai, Junjie [1 ]
Yang, Yuekui [1 ]
Wang, Di [1 ]
机构
[1] Tencent AI Lab, Shenzhen, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Although attention-based Neural Machine Translation (NMT) has achieved remarkable progress in recent years, it still suffers from issues of repeating and dropping translations. To alleviate these issues, we propose a novel key-value memory-augmented attention model for NMT, called KVMEMATT. Specifically, we maintain a timely updated key-memory to keep track of attention history and a fixed value-memory to store the representation of source sentence throughout the whole translation process. Via nontrivial transformations and iterative interactions between the two memories, the decoder focuses on more appropriate source word(s) for predicting the next target word at each decoding step, therefore can improve the adequacy of translations. Experimental results on Chinese double right arrow English and WMT17 German double left right arrow English translation tasks demonstrate the superiority of the proposed model.
引用
收藏
页码:2574 / 2580
页数:7
相关论文
共 50 条
  • [1] Generalized Key-Value Memory to Flexibly Adjust Redundancy in Memory-Augmented Networks
    Kleyko, Denis
    Karunaratne, Geethan
    Rabaey, Jan M.
    Sebastian, Abu
    Rahimi, Abbas
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (12) : 10993 - 10998
  • [2] Memory-augmented Chinese-Uyghur Neural Machine Translation
    Zhang, Shiyue
    Mahmut, Gulnigar
    Wang, Dong
    Hamdulla, Askar
    2017 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC 2017), 2017, : 1092 - 1096
  • [3] Rethinking Translation Memory Augmented Neural Machine Translation
    Hao, Hongkun
    Huang, Guoping
    Liu, Lemao
    Zhang, Zhirui
    Shi, Shuming
    Wang, Rui
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 2589 - 2605
  • [4] Quantized Memory-Augmented Neural Networks
    Park, Seongsik
    Kim, Seijoon
    Lee, Seil
    Bae, Ho
    Yoon, Sungroh
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 3909 - 3916
  • [5] Memory-Augmented Episodic Value Network
    Zeng, Fanyu
    Xing, Guangyu
    Han, Guang
    2022 IEEE CONFERENCE ON GAMES, COG, 2022, : 315 - 321
  • [6] Memory-Augmented Attention Network for Sequential Recommendation
    Hu, Cheng
    He, Peijian
    Sha, Chaofeng
    Niu, Junyu
    WEB INFORMATION SYSTEMS ENGINEERING - WISE 2019, 2019, 11881 : 228 - 242
  • [7] Manna: An Accelerator for Memory-Augmented Neural Networks
    Stevens, Jacob R.
    Ranjan, Ashish
    Das, Dipankar
    Kaul, Bharat
    Raghunathan, Anand
    MICRO'52: THE 52ND ANNUAL IEEE/ACM INTERNATIONAL SYMPOSIUM ON MICROARCHITECTURE, 2019, : 794 - 806
  • [8] A Memory-Augmented Neural Model for Automated Grading
    Zhao, Siyuan
    Zhang, Yaqiong
    Xiong, Xiaolu
    Botelho, Anthony
    Heffernan, Neil
    PROCEEDINGS OF THE FOURTH (2017) ACM CONFERENCE ON LEARNING @ SCALE (L@S'17), 2017, : 189 - 192
  • [9] Memory-Augmented Attention Model for Scene Text Recognition
    Wang, Cong
    Yin, Fei
    Liu, Cheng-Lin
    PROCEEDINGS 2018 16TH INTERNATIONAL CONFERENCE ON FRONTIERS IN HANDWRITING RECOGNITION (ICFHR), 2018, : 62 - 67
  • [10] Memory-augmented Dynamic Neural Relational Inference
    Gong, Dong
    Zhang, Zhen
    Shi, Javen Qinfeng
    Van den Hengel, Anton
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 11823 - 11832