The Unreasonable Volatility of Neural Machine Translation Models

被引:0
|
作者
Fadaee, Marzieh [1 ,2 ]
Monz, Christof [1 ]
机构
[1] Univ Amsterdam, Informat Inst, Amsterdam, Netherlands
[2] Zeta Alpha Vector, Amsterdam, Netherlands
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent works have shown that while Neural Machine Translation (NMT) models achieve impressive performance, questions about understanding the behaviour of these models remain unanswered. We investigate the unexpected volatility of NMT models where the input is semantically and syntactically correct. We discover that with trivial modifications of source sentences, we can identify cases where unexpected changes happen in the translation and in the worst case lead to mistranslations. This volatile behaviour of translating extremely similar sentences in surprisingly different ways highlights the underlying generalization problem of current NMT models. We find that both RNN and Transformer models display volatile behaviour in 26% and 19% of sentence variations, respectively.
引用
收藏
页码:88 / 96
页数:9
相关论文
共 50 条
  • [1] On the Sparsity of Neural Machine Translation Models
    Wang, Yong
    Wang, Longyue
    Li, Victor O. K.
    Tu, Zhaopeng
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 1060 - 1066
  • [2] Compact Personalized Models for Neural Machine Translation
    Wuebker, Joern
    Simianer, Patrick
    DeNero, John
    2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 881 - 886
  • [3] Machine Translation based on Neural Network Language Models
    Zamora-Martinez, Francisco
    Jose Castro-Bleda, Maria
    PROCESAMIENTO DEL LENGUAJE NATURAL, 2010, (45): : 221 - 228
  • [4] On the linguistic representational power of neural machine translation models
    Belinkov Y.
    Durrani N.
    Dalvi F.
    Sajjad H.
    Glass J.
    1600, MIT Press Journals (46): : 1 - 52
  • [5] Multiscale Collaborative Deep Models for Neural Machine Translation
    Wei, Xiangpeng
    Yu, Heng
    Hu, Yue
    Zhang, Yue
    Weng, Rongxiang
    Luo, Weihua
    58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), 2020, : 414 - 426
  • [6] On the Linguistic Representational Power of Neural Machine Translation Models
    Belinkov, Yonatan
    Durrani, Nadir
    Dalvi, Fahim
    Sajjad, Hassan
    Glass, James
    COMPUTATIONAL LINGUISTICS, 2020, 46 (01) : 1 - 52
  • [7] Improving Neural Machine Translation Models with Monolingual Data
    Sennrich, Rico
    Haddow, Barry
    Birch, Alexandra
    PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, 2016, : 86 - 96
  • [8] An Investigation on Statistical Machine Translation with Neural Language Models
    Zhao, Yinggong
    Huang, Shujian
    Chen, Huadong
    Chen, Jiajun
    CHINESE COMPUTATIONAL LINGUISTICS AND NATURAL LANGUAGE PROCESSING BASED ON NATURALLY ANNOTATED BIG DATA, CCL 2014, 2014, 8801 : 175 - 186
  • [9] Lattice-to-sequence attentional Neural Machine Translation models
    Tan, Zhixing
    Su, Jinsong
    Wang, Boli
    Chen, Yidong
    Shi, Xiaodong
    NEUROCOMPUTING, 2018, 284 : 138 - 147
  • [10] Training Deeper Neural Machine Translation Models with Transparent Attention
    Bapna, Ankur
    Chen, Mia Xu
    Firat, Orhan
    Cao, Yuan
    Wu, Yonghui
    2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 3028 - 3033