Neural Machine Translation with Key-Value Memory-Augmented Attention

被引:0
|
作者
Meng, Fandong [1 ]
Tu, Zhaopeng [1 ]
Cheng, Yong [1 ]
Wu, Haiyang [1 ]
Zhai, Junjie [1 ]
Yang, Yuekui [1 ]
Wang, Di [1 ]
机构
[1] Tencent AI Lab, Shenzhen, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Although attention-based Neural Machine Translation (NMT) has achieved remarkable progress in recent years, it still suffers from issues of repeating and dropping translations. To alleviate these issues, we propose a novel key-value memory-augmented attention model for NMT, called KVMEMATT. Specifically, we maintain a timely updated key-memory to keep track of attention history and a fixed value-memory to store the representation of source sentence throughout the whole translation process. Via nontrivial transformations and iterative interactions between the two memories, the decoder focuses on more appropriate source word(s) for predicting the next target word at each decoding step, therefore can improve the adequacy of translations. Experimental results on Chinese double right arrow English and WMT17 German double left right arrow English translation tasks demonstrate the superiority of the proposed model.
引用
收藏
页码:2574 / 2580
页数:7
相关论文
共 50 条
  • [41] Memory-Augmented Graph Neural Networks: A Brain-Inspired Review
    Ma G.
    Vo V.A.
    Willke T.L.
    Ahmed N.K.
    IEEE Transactions on Artificial Intelligence, 5 (05): : 2011 - 2025
  • [42] Learning a Key-Value Memory Co-Attention Matching Network for Person Re-Identification
    Zhang, Yaqing
    Li, Xi
    Zhang, Zhongfei
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 9235 - 9242
  • [43] Recurrent Attention for Neural Machine Translation
    Zeng, Jiali
    Wu, Shuangzhi
    Yin, Yongjing
    Jiang, Yufan
    Li, Mu
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 3216 - 3225
  • [44] Neural Machine Translation with Deep Attention
    Zhang, Biao
    Xiong, Deyi
    Su, Jinsong
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2020, 42 (01) : 154 - 163
  • [45] Concept learning through deep reinforcement learning with memory-augmented neural networks
    Shi, Jing
    Xu, Jiaming
    Yao, Yiqun
    Xu, Bo
    NEURAL NETWORKS, 2019, 110 : 47 - 54
  • [46] Energy-Efficient Inference Accelerator for Memory-Augmented Neural Networks on an FPGA
    Park, Seongsik
    Jang, Jaehee
    Kim, Seijoon
    Yoon, Sungroh
    2019 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE), 2019, : 1587 - 1590
  • [47] In-Memory Key-Value Store Live Migration with NetMigrate
    Zhu, Zeying
    Zhao, Yibo
    Liu, Zaoxing
    PROCEEDINGS OF THE 21ST USENIX SYMPOSIUM ON NETWORKED SYSTEMS DESIGN AND IMPLEMENTATION, NSDI 24, 2024, : 209 - 224
  • [48] InK: In-Kernel Key-Value Storage with Persistent Memory
    Ha, Minjong
    Kim, Sang-Hoon
    ELECTRONICS, 2020, 9 (11) : 1 - 22
  • [49] Attention-via-Attention Neural Machine Translation
    Zhao, Shenjian
    Zhang, Zhihua
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 563 - 570
  • [50] In-Memory Key-Value Store Live Migration with NetMigrate
    Zhu, Zeying
    Zhao, Yibo
    Liu, Zaoxing
    PROCEEDINGS OF THE 22ND USENIX CONFERENCE ON FILE AND STORAGE TECHNOLOGIES, FAST 24, 2024, : 209 - 224