Neural Machine Translation with Key-Value Memory-Augmented Attention

被引:0
|
作者
Meng, Fandong [1 ]
Tu, Zhaopeng [1 ]
Cheng, Yong [1 ]
Wu, Haiyang [1 ]
Zhai, Junjie [1 ]
Yang, Yuekui [1 ]
Wang, Di [1 ]
机构
[1] Tencent AI Lab, Shenzhen, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Although attention-based Neural Machine Translation (NMT) has achieved remarkable progress in recent years, it still suffers from issues of repeating and dropping translations. To alleviate these issues, we propose a novel key-value memory-augmented attention model for NMT, called KVMEMATT. Specifically, we maintain a timely updated key-memory to keep track of attention history and a fixed value-memory to store the representation of source sentence throughout the whole translation process. Via nontrivial transformations and iterative interactions between the two memories, the decoder focuses on more appropriate source word(s) for predicting the next target word at each decoding step, therefore can improve the adequacy of translations. Experimental results on Chinese double right arrow English and WMT17 German double left right arrow English translation tasks demonstrate the superiority of the proposed model.
引用
收藏
页码:2574 / 2580
页数:7
相关论文
共 50 条
  • [21] A Memory-augmented Conditional Neural Process model for traffic prediction
    Wei, Ye
    He, Haitao
    Yuan, Kunhao
    Schaefer, Gerald
    Ji, Zhigang
    Hui, Fang
    KNOWLEDGE-BASED SYSTEMS, 2024, 304
  • [22] Biological learning in key-value memory networks
    Tyulmankov, Danil
    Fang, Ching
    Vadaparty, Annapurna
    Yang, Guangyu Robert
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [23] Attention and Memory-Augmented Networks for Dual-View Sequential Learning
    He, Yong
    Wang, Cheng
    Li, Nan
    Zeng, Zhenyu
    KDD '20: PROCEEDINGS OF THE 26TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2020, : 125 - 134
  • [24] Robust high-dimensional memory-augmented neural networks
    Geethan Karunaratne
    Manuel Schmuck
    Manuel Le Gallo
    Giovanni Cherubini
    Luca Benini
    Abu Sebastian
    Abbas Rahimi
    Nature Communications, 12
  • [25] Scaling Memory-Augmented Neural Networks with Sparse Reads and Writes
    Rae, Jack W.
    Hunt, Jonathan J.
    Harley, Tim
    Danihelka, Ivo
    Senior, Andrew
    Wayne, Greg
    Graves, Alex
    Lillicrap, Timothy P.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29
  • [26] Attention-Augmented Machine Memory
    Xin Lin
    Guoqiang Zhong
    Kang Chen
    Qingyang Li
    Kaizhu Huang
    Cognitive Computation, 2021, 13 : 751 - 760
  • [27] Attention-Augmented Machine Memory
    Lin, Xin
    Zhong, Guoqiang
    Chen, Kang
    Li, Qingyang
    Huang, Kaizhu
    COGNITIVE COMPUTATION, 2021, 13 (03) : 751 - 760
  • [28] MEFET-Based CAM/TCAM for Memory-Augmented Neural Networks
    Sanjeet, Sai
    Bird, Jonathan
    Sahoo, Bibhu Datta
    IEEE JOURNAL ON EXPLORATORY SOLID-STATE COMPUTATIONAL DEVICES AND CIRCUITS, 2024, 10 : 31 - 39
  • [29] Dynamic Key-Value Memory Networks for Knowledge Tracing
    Zhang, Jiani
    Shi, Xingjian
    King, Irwin
    Yeung, Dit-Yan
    PROCEEDINGS OF THE 26TH INTERNATIONAL CONFERENCE ON WORLD WIDE WEB (WWW'17), 2017, : 765 - 774
  • [30] Lightweight and Accurate Memory Allocation in Key-Value Cache
    Pan, Cheng
    Zhou, Lan
    Luo, Yingwei
    Wang, Xiaolin
    Wang, Zhenlin
    INTERNATIONAL JOURNAL OF PARALLEL PROGRAMMING, 2019, 47 (03) : 451 - 466