Named Entity Correction in Neural Machine Translation Using the Attention Alignment Map

被引:4
|
作者
Lee, Jangwon [1 ,2 ]
Lee, Jungi [3 ]
Lee, Minho [3 ]
Jang, Gil-Jin [2 ,4 ]
机构
[1] SK Holdings C&C, Suwon 13558, South Korea
[2] Kyungpook Natl Univ, Sch Elect & Elect Engn, Daegu 41566, South Korea
[3] Kyungpook Natl Univ, Dept Artificial Intelligence, Daegu 41566, South Korea
[4] Kyungpook Natl Univ, Sch Elect Engn, Daegu 41566, South Korea
来源
APPLIED SCIENCES-BASEL | 2021年 / 11卷 / 15期
基金
新加坡国家研究基金会;
关键词
neural networks; recurrent neural networks; natural language processing; neural machine translation; named entity recognition; SEQUENCE;
D O I
10.3390/app11157026
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Featured Application machine translation; information retrieval; text-to-speech. Neural machine translation (NMT) methods based on various artificial neural network models have shown remarkable performance in diverse tasks and have become mainstream for machine translation currently. Despite the recent successes of NMT applications, a predefined vocabulary is still required, meaning that it cannot cope with out-of-vocabulary (OOV) or rarely occurring words. In this paper, we propose a postprocessing method for correcting machine translation outputs using a named entity recognition (NER) model to overcome the problem of OOV words in NMT tasks. We use attention alignment mapping (AAM) between the named entities of input and output sentences, and mistranslated named entities are corrected using word look-up tables. The proposed method corrects named entities only, so it does not require retraining of existing NMT models. We carried out translation experiments on a Chinese-to-Korean translation task for Korean historical documents, and the evaluation results demonstrated that the proposed method improved the bilingual evaluation understudy (BLEU) score by 3.70 from the baseline.
引用
收藏
页数:21
相关论文
共 50 条
  • [41] Neural machine translation for Indian language pair using hybrid attention mechanism
    Basab Nath
    Sunita Sarkar
    Surajeet Das
    Somnath Mukhopadhyay
    Innovations in Systems and Software Engineering, 2024, 20 : 175 - 183
  • [42] Enhancement of Encoder and Attention Using Target Monolingual Corpora in Neural Machine Translation
    Imamura, Kenji
    Fujita, Atsushi
    Sumita, Eiichiro
    NEURAL MACHINE TRANSLATION AND GENERATION, 2018, : 55 - 63
  • [43] A Named Entity Recognition System for Malayalam using Neural Networks
    Ajees, A. P.
    Idicula, Sumam Mary
    8TH INTERNATIONAL CONFERENCE ON ADVANCES IN COMPUTING & COMMUNICATIONS (ICACC-2018), 2018, 143 : 962 - 969
  • [44] Neural machine translation for Indian language pair using hybrid attention mechanism
    Nath, Basab
    Sarkar, Sunita
    Das, Surajeet
    Mukhopadhyay, Somnath
    INNOVATIONS IN SYSTEMS AND SOFTWARE ENGINEERING, 2024, 20 (02) : 175 - 183
  • [45] A self-attention based neural architecture for Chinese medical named entity recognition
    Wan, Qian
    Liu, Jie
    Wei, Luona
    Ji, Bin
    MATHEMATICAL BIOSCIENCES AND ENGINEERING, 2020, 17 (04) : 3498 - 3511
  • [46] Attention With Sparsity Regularization for Neural Machine Translation and Summarization
    Zhang, Jiajun
    Zhao, Yang
    Li, Haoran
    Zong, Chengqing
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2019, 27 (03) : 507 - 518
  • [47] Neural Machine Translation with Target-Attention Model
    Yang, Mingming
    Zhang, Min
    Chen, Kehai
    Wang, Rui
    Zhao, Tiejun
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2020, E103D (03) : 684 - 694
  • [48] Syntax-Directed Attention for Neural Machine Translation
    Chen, Kehai
    Wang, Rui
    Utiyama, Masao
    Sumita, Eiichiro
    Zhao, Tiejun
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 4792 - 4799
  • [49] Dynamic Attention Aggregation with BERT for Neural Machine Translation
    Zhang, JiaRui
    Li, HongZheng
    Shi, ShuMin
    Huang, HeYan
    Hu, Yue
    Wei, XiangPeng
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [50] Synchronous Syntactic Attention for Transformer Neural Machine Translation
    Deguchi, Hiroyuki
    Tamura, Akihiro
    Ninomiya, Takashi
    ACL-IJCNLP 2021: THE 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING: PROCEEDINGS OF THE STUDENT RESEARCH WORKSHOP, 2021, : 348 - 355