KinyaBERT: a Morphology-aware Kinyarwanda Language Model

被引:0
|
作者
Nzeyimana, Antoine [1 ]
Rubungo, Andre Niyongabo [2 ]
机构
[1] Univ Massachusetts, Amherst, MA 01003 USA
[2] Univ Politecn Cataluna, Barcelona, Spain
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pre-trained language models such as BERT have been successful at tackling many natural language processing tasks. However, the unsupervised sub-word tokenization methods commonly used in these models (e.g., byte-pair encoding - BPE) are sub-optimal at handling morphologically rich languages. Even given a morphological analyzer, naive sequencing of morphemes into a standard BERT architecture is inefficient at capturing morphological compositionality and expressing word-relative syntactic regularities. We address these challenges by proposing a simple yet effective twotier BERT architecture that leverages a morphological analyzer and explicitly represents morphological compositionality. Despite the success of BERT, most of its evaluations have been conducted on high-resource languages, obscuring its applicability on low-resource languages. We evaluate our proposed method on the low-resource morphologically rich Kinyarwanda language, naming the proposed model architecture KinyaBERT. A robust set of experimental results reveal that KinyaBERT outperforms solid baselines by 2% in F1 score on a named entity recognition task and by 4.3% in average score of a machine-translated GLUE benchmark. KinyaBERT fine-tuning has better convergence and achieves more robust results on multiple tasks even in the presence of translation noise.(1)
引用
收藏
页码:5347 / 5363
页数:17
相关论文
共 50 条
  • [41] DagoBERT: Generating Derivational Morphology with a Pretrained Language Model
    Hofmannt, Valentin
    Pierrehumbertt, Janet B.
    Schiitzet, Hinrich
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 3848 - 3861
  • [42] A Formal Model and Composition Language for Context-Aware Service Protocols
    Cubo, Javier
    Canal, Carlos
    Pimentel, Ernesto
    Salauen, Gwen
    CASTA 2009: WORKSHOP ON CONTEXT-AWARE SOFTWARE TECHNOLOGY AND APPLICATIONS, 2009, : 17 - 20
  • [43] PAPEL: A Language and Model for Provenance-Aware Policy Definition and Execution
    Ringelstein, Christoph
    Staab, Steffen
    BUSINESS PROCESS MANAGEMENT, 2010, 6336 : 195 - 210
  • [44] Empower Sequence Labeling with Task-Aware Neural Language Model
    Liu, Liyuan
    Shang, Jingbo
    Ren, Xiang
    Xu, Frank F.
    Gui, Huan
    Peng, Jian
    Han, Jiawei
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 5253 - 5260
  • [45] Retrofitting Structure-aware Transformer Language Model for End Tasks
    Fei, Hao
    Ren, Yafeng
    Ji, Donghong
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 2151 - 2161
  • [46] EFFICIENTLY-REALIZED MODEL OF THE NATURAL-LANGUAGE MORPHOLOGY
    GELBUKH, AF
    NAUCHNO-TEKHNICHESKAYA INFORMATSIYA SERIYA 2-INFORMATSIONNYE PROTSESSY I SISTEMY, 1992, (01): : 24 - 31
  • [47] Topic Structure-Aware Neural Language Model: Unified language model that maintains word and topic ordering by their embedded representations
    Kawamae, Noriaki
    WEB CONFERENCE 2019: PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE (WWW 2019), 2019, : 2900 - 2906
  • [48] Domain-Aware Word Segmentation for Chinese Language: A Document-Level Context-Aware Model
    Huang, Kaiyu
    Xiao, Keli
    Mo, Fengran
    Jin, Bo
    Liu, Zhuang
    Huang, Degen
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2022, 21 (02)
  • [49] An Emotion-Aware Human-Computer Negotiation Model Powered by Pretrained Language Model
    Luo, Xudong
    Deng, Zhiqi
    Sun, Kaili
    Lin, Pingping
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT IV, KSEM 2024, 2024, 14887 : 243 - 259
  • [50] Textual Differential Privacy for Context-Aware Reasoning with Large Language Model
    Yu, Junwei
    Zhou, Jieyu
    Ding, Yepeng
    Zhang, Lingfeng
    Guo, Yuheng
    Sato, Hiroyuki
    2024 IEEE 48TH ANNUAL COMPUTERS, SOFTWARE, AND APPLICATIONS CONFERENCE, COMPSAC 2024, 2024, : 988 - 997