English-Assamese neural machine translation using prior alignment and pre-trained language model

被引:9
|
作者
Laskar, Sahinur Rahman [1 ]
Paul, Bishwaraj [1 ]
Dadure, Pankaj [2 ]
Manna, Riyanka [3 ]
Pakray, Partha [1 ]
Bandyopadhyay, Sivaji [1 ]
机构
[1] Natl Inst Technol, Dept Comp Sci & Engn, Silchar 788010, Assam, India
[2] Univ Petr & Energy Studies, Sch Comp Sci, Dehra Dun 248007, Uttarakhand, India
[3] Jadavpur Univ, Dept Comp Sci & Engn, Kolkata 700032, W Bengal, India
来源
关键词
Low-resource; NMT; English-Assamese; Alignment; Language model;
D O I
10.1016/j.csl.2023.101524
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In a multilingual country like India, automatic natural language translation plays a key role in building a community with different linguistic people. Many researchers have explored and improved the translation process for high-resource languages such as English, German, etc., and achieved state-of-the-art results. However, the unavailability of adequate data is the prime obstacle to automatic natural language translation of low-resource north-eastern Indian languages such as Mizo, Khasi, and Assamese. Though the recent past has witnessed a deluge in several automatic natural language translation systems for low-resource languages, the low values of their evaluation measures indicate the scope for improvement. In the recent past, the neural machine translation approach has significantly improved translation quality, and the credit goes to the availability of a huge amount of data. Subsequently, the neural machine translation approach for low-resource language is underrepresented due to the unavailability of adequate data. In this work, we have considered a low-resource English-Assamese pair using the transformer-based neural machine translation, which leverages the use of prior alignment and a pre-trained language model. To extract alignment information from the source-target sentences, we have used the pre-trained multilingual contextual embeddings-based alignment technique. Also, the transformer-based language model is built using monolingual target sentences. With the use of both prior alignment and a pre-trained language model, the transformer-based neural machine translation model shows improvement, and we have achieved state-of-the-art results for the English-to-Assamese and Assamese-to-English translation, respectively.
引用
收藏
页数:16
相关论文
共 50 条
  • [21] Machine Unlearning of Pre-trained Large Language Models
    Yao, Jin
    Chien, Eli
    Du, Minxin
    Niu, Xinyao
    Wang, Tianhao
    Cheng, Zezhou
    Yue, Xiang
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 8403 - 8419
  • [22] Adder Encoder for Pre-trained Language Model
    Ding, Jianbang
    Zhang, Suiyun
    Li, Linlin
    CHINESE COMPUTATIONAL LINGUISTICS, CCL 2023, 2023, 14232 : 339 - 347
  • [23] Multilingual Translation via Grafting Pre-trained Language Models
    Sun, Zewei
    Wang, Mingxuan
    Li, Lei
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 2735 - 2747
  • [24] Pre-trained language model for code-mixed text in Indonesian, Javanese, and English using transformer
    Ahmad Fathan Hidayatullah
    Rosyzie Anna Apong
    Daphne Teck Ching Lai
    Atika Qazi
    Social Network Analysis and Mining, 15 (1)
  • [25] Modeling Second Language Acquisition with pre-trained neural language models
    Palenzuela, Alvaro J. Jimenez
    Frasincar, Flavius
    Trusca, Maria Mihaela
    EXPERT SYSTEMS WITH APPLICATIONS, 2022, 207
  • [26] Automatic Fixation of Decompilation Quirks Using Pre-trained Language Model
    Kaichi, Ryunosuke
    Matsumoto, Shinsuke
    Kusumoto, Shinji
    PRODUCT-FOCUSED SOFTWARE PROCESS IMPROVEMENT, PROFES 2023, PT I, 2024, 14483 : 259 - 266
  • [27] Surgicberta: a pre-trained language model for procedural surgical language
    Bombieri, Marco
    Rospocher, Marco
    Ponzetto, Simone Paolo
    Fiorini, Paolo
    INTERNATIONAL JOURNAL OF DATA SCIENCE AND ANALYTICS, 2024, 18 (01) : 69 - 81
  • [28] Improving Braille-Chinese translation with jointly trained and pre-trained language models
    Huang, Tianyuan
    Su, Wei
    Liu, Lei
    Cai, Chuan
    Yu, Hailong
    Yuan, Yongna
    DISPLAYS, 2024, 82
  • [29] ZeroEA: A Zero-Training Entity Alignment Framework via Pre-Trained Language Model
    Huo, Nan
    Cheng, Reynold
    Kao, Ben
    Ning, Wentao
    Haldar, Nur Al Hasan
    Li, Xiaodong
    Li, Jinyang
    Najafi, Mohammad Matin
    Li, Tian
    Qu, Ge
    PROCEEDINGS OF THE VLDB ENDOWMENT, 2024, 17 (07): : 1765 - 1774
  • [30] Pre-trained Language Model for Biomedical Question Answering
    Yoon, Wonjin
    Lee, Jinhyuk
    Kim, Donghyeon
    Jeong, Minbyul
    Kang, Jaewoo
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2019, PT II, 2020, 1168 : 727 - 740