On the Cross-lingual Transferability of Monolingual Representations

被引:0
|
作者
Artetxe, Mikel [1 ,2 ]
Ruder, Sebastian [2 ]
Yogatama, Dani [2 ]
机构
[1] Univ Basque Country UPV EHU, HiTZ Ctr, Leioa, Spain
[2] DeepMind, London, England
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
State-of-the-art unsupervised multilingual models (e.g., multilingual BERT) have been shown to generalize in a zero-shot cross-lingual setting. This generalization ability has been attributed to the use of a shared subword vocabulary and joint training across multiple languages giving rise to deep multilingual abstractions. We evaluate this hypothesis by designing an alternative approach that transfers a monolingual model to new languages at the lexical level. More concretely, we first train a transformer-based masked language model on one language, and transfer it to a new language by learning a new embedding matrix with the same masked language modeling objective-freezing parameters of all other layers. This approach does not rely on a shared vocabulary or joint training. However, we show that it is competitive with multilingual BERT on standard cross-lingual classification benchmarks and on a new Cross-lingual Question Answering Dataset (XQuAD). Our results contradict common beliefs of the basis of the generalization ability of multilingual models and suggest that deep monolingual models learn some abstractions that generalize across languages. We also release XQuAD as a more comprehensive cross-lingual benchmark, which comprises 240 paragraphs and 1190 question-answer pairs from SQuAD v1.1 translated into ten languages by professional translators.
引用
收藏
页码:4623 / 4637
页数:15
相关论文
共 50 条
  • [1] Cross-lingual Transfer of Monolingual Models
    Gogoulou, Evangelia
    Ekgren, Ariel
    Isbister, Tim
    Sahlgren, Magnus
    LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 948 - 955
  • [2] BERT for Monolingual and Cross-Lingual Reverse Dictionary
    Yan, Hang
    Li, Xiaonan
    Qiu, Xipeng
    Deng, Bocao
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 4329 - 4338
  • [3] Monolingual, multilingual and cross-lingual code comment classification
    Kostic, Marija
    Batanovic, Vuk
    Nikolic, Bosko
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 124
  • [4] Monolingual and Cross-Lingual Knowledge Transfer for Topic Classification
    D. Karpov
    M. Burtsev
    Journal of Mathematical Sciences, 2024, 285 (1) : 36 - 48
  • [5] Cross-Lingual Learning with Distributed Representations
    Pikuliak, Matus
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 8032 - 8033
  • [6] Reproducing Monolingual, Multilingual and Cross-Lingual CEFR Predictions
    Bestgen, Yves
    PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2020), 2020, : 5595 - 5602
  • [7] Monolingual and Cross-Lingual Acceptability Judgments with the Italian CoLA corpus
    Trotta, Daniela
    Guarasci, Raffaele
    Leonardelli, Elisa
    Tonelli, Sara
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 2929 - 2940
  • [8] Can Monolingual Pretrained Models Help Cross-Lingual Classification?
    Chi, Zewen
    Dong, Li
    Wei, Furu
    Mao, Xian-Ling
    Huang, Heyan
    1ST CONFERENCE OF THE ASIA-PACIFIC CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 10TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (AACL-IJCNLP 2020), 2020, : 12 - 17
  • [9] Cross-lingual Evidence Improves Monolingual Fake News Detection
    Dementieva, Daryna
    Panchenko, Alexander
    ACL-IJCNLP 2021: THE 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING: PROCEEDINGS OF THE STUDENT RESEARCH WORKSHOP, 2021, : 310 - 320
  • [10] Contextualized Embeddings Encode Monolingual and Cross-lingual Knowledge of Idiomaticity
    Fakharian, Samin
    Cook, Paul
    MWE 2021: THE 17TH WORKSHOP ON MULTIWORD EXPRESSIONS, 2021, : 23 - 32