Improving Adversarial Neural Machine Translation for Morphologically Rich Language

被引:10
|
作者
Mi, Chenggang [1 ]
Xie, Lei [1 ]
Zhang, Yanning [1 ]
机构
[1] Northwestern Polytech Univ, Sch Comp Sci, Natl Engn Lab Integrated Aerosp Ground Ocean Big, Xian 710072, Peoples R China
基金
中国国家自然科学基金;
关键词
Neural machine translation (NMT); morphologically rich language; adversarial training; morphological word embedding; multiple references;
D O I
10.1109/TETCI.2019.2960546
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Generative adversarial networks (GAN) have great successes on natural language processing (NLP) and neural machine translation (NMT). However, the existing discriminator in GAN for NMT only combines two words as one query to train the translation models, which restrict the discriminator to be more meaningful and fail to apply rich monolingual information. Recent studies only consider one single reference translation during model training, this limit the GAN model to learn sufficient information about the representation of source sentence. These situations are even worse when languages are morphologically rich. In this article, an extended version of GAN model for neural machine translation is proposed to optimize the performance of morphologically rich language translation. In particular, we use the morphological word embedding instead of word embedding as input in GAN model to enrich the representation of words and overcome the data sparsity problem during model training. Moreover, multiple references are integrated into discriminator to make the model consider more context information and adapt to the diversity of different languages. Experimental results on German <-> English, French <-> English, Czech <-> English, Finnish <-> English, Turkish <-> English, Chinese <-> English, Finnish <-> Turkish and Turkish <-> Czech translation tasks demonstrate that our method achieves significant improvements over baseline systems.
引用
收藏
页码:417 / 426
页数:10
相关论文
共 50 条
  • [41] Improving Machine Translation Formality with Large Language Models
    Yang, Murun
    Li, Fuxue
    CMC-COMPUTERS MATERIALS & CONTINUA, 2025, 82 (02): : 2061 - 2075
  • [42] INTRAWORD DECOMPOSITION IN A MORPHOLOGICALLY RICH LANGUAGE
    NIEMI, J
    LAINE, M
    KOIVUSELKASALLINEN, P
    INTERNATIONAL JOURNAL OF PSYCHOLOGY, 1992, 27 (3-4) : 86 - 86
  • [43] Improving sentiment analysis performance on morphologically rich languages: Language and domain independent approach
    Kincl, Tomas
    Novak, Michal
    Pribil, Jiri
    COMPUTER SPEECH AND LANGUAGE, 2019, 56 : 36 - 51
  • [44] Improving Real-time Recognition of Morphologically Rich Speech with Transformer Language Model
    Tarjan, Balazs
    Szaszak, Gyorgy
    Fegyo, Tibor
    Mihajlik, Peter
    2020 11TH IEEE INTERNATIONAL CONFERENCE ON COGNITIVE INFOCOMMUNICATIONS (COGINFOCOM 2020), 2020, : 491 - 495
  • [45] Improving Neural Machine Translation with AMR Semantic Graphs
    Nguyen, Long H. B.
    Pham, Viet H.
    Dinh, Dien
    MATHEMATICAL PROBLEMS IN ENGINEERING, 2021, 2021
  • [46] Improving neural machine translation with latent features feedback
    Li, Yachao
    Li, Junhui
    Zhang, Min
    NEUROCOMPUTING, 2021, 463 : 368 - 378
  • [47] Improving neural machine translation with sentence alignment learning
    Shi, Xuewen
    Huang, Heyan
    Jian, Ping
    Tang, Yi-Kun
    NEUROCOMPUTING, 2021, 420 : 15 - 26
  • [48] Improving Neural Machine Translation Models with Monolingual Data
    Sennrich, Rico
    Haddow, Barry
    Birch, Alexandra
    PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, 2016, : 86 - 96
  • [49] Measuring and Improving Faithfulness of Attention in Neural Machine Translation
    Moradi, Pooya
    Kambhatla, Nishant
    Sarkar, Anoop
    16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 2791 - 2802
  • [50] Understanding and Improving Hidden Representation for Neural Machine Translation
    Li, Guanlin
    Liu, Lemao
    Li, Xintong
    Zhu, Conghui
    Zhao, Tiejun
    Shi, Shuming
    2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, 2019, : 466 - 477