Language Model Priming for Cross-Lingual Event Extraction

被引:0
|
作者
Fincke, Steven [1 ]
Agarwal, Shantanu [1 ]
Miller, Scott [1 ]
Boschee, Elizabeth [1 ]
机构
[1] Univ Southern Calif, Inst Informat Sci, Los Angeles, CA 90007 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present a novel, language-agnostic approach to "priming" language models for the task of event extraction, providing particularly effective performance in low-resource and zero-shot cross-lingual settings. With priming, we augment the input to the transformer stack's language model differently depending on the question(s) being asked of the model at runtime. For instance, if the model is being asked to identify arguments for the trigger protested, we will provide that trigger as part of the input to the language model, allowing it to produce different representations for candidate arguments than when it is asked about arguments for the trigger arrest elsewhere in the same sentence. We show that by enabling the language model to better compensate for the deficits of sparse and noisy training data, our approach improves both trigger and argument detection and classification significantly over the state of the art in a zero-shot cross-lingual setting.
引用
收藏
页码:10627 / 10635
页数:9
相关论文
共 50 条
  • [1] Cross-lingual Language Model Pretraining
    Conneau, Alexis
    Lample, Guillaume
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [2] Cross-lingual Structure Transfer for Relation and Event Extraction
    Subburathinam, Ananya
    Lu, Di
    Ji, Heng
    May, Jonathan
    Chang, Shih-Fu
    Sil, Avirup
    Voss, Clare
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 313 - 325
  • [3] Language Anisotropic Cross-Lingual Model Editing
    Xu, Yang
    Hou, Yutai
    Che, Wanxiang
    Zhang, Min
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 5554 - 5569
  • [4] Cross-lingual Language Model Pretraining for Retrieval
    Yu, Puxuan
    Fei, Hongliang
    Li, Ping
    PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE 2021 (WWW 2021), 2021, : 1029 - 1039
  • [5] Multilingual Generative Language Models for Zero-Shot Cross-Lingual Event Argument Extraction
    Huang, Kuan-Hao
    Hsu, I-Hung
    Natarajan, Premkumar
    Chang, Kai-Wei
    Peng, Nanyun
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 4633 - 4646
  • [6] Cross-lingual Structure Transfer for Zero-resource Event Extraction
    Lu, Di
    Subburathinam, Ananya
    Ji, Heng
    May, Jonathan
    Chang, Shih-Fu
    Sil, Avirup
    Voss, Clare
    PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2020), 2020, : 1976 - 1981
  • [7] Multi-lingual and Cross-lingual timeline extraction
    Laparra, Egoitz
    Agerri, Rodrigo
    Aldabe, Itziar
    Rigau, German
    KNOWLEDGE-BASED SYSTEMS, 2017, 133 : 77 - 89
  • [8] Cross-Lingual Latent Topic Extraction
    Zhang, Duo
    Mei, Qiaozhu
    Zhai, ChengXiang
    ACL 2010: 48TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2010, : 1128 - 1137
  • [9] Zero-Shot Cross-Lingual Event Argument Extraction with Language-Oriented Prefix-Tuning
    Cao, Pengfei
    Jin, Zhuoran
    Chen, Yubo
    Liu, Kang
    Zhao, Jun
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 11, 2023, : 12589 - 12597
  • [10] Unsupervised Domain Adaptation of a Pretrained Cross-Lingual Language Model
    Li, Juntao
    He, Ruidan
    Ye, Hai
    Ng, Hwee Tou
    Bing, Lidong
    Yan, Rui
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 3672 - 3678