PTEKC: pre-training with event knowledge of ConceptNet for cross-lingual event causality identification

被引:0
|
作者
Zhu, Enchang [1 ,2 ]
Yu, Zhengtao [1 ,2 ]
Huang, Yuxin [1 ,2 ]
Gao, Shengxiang [1 ,2 ]
Xian, Yantuan [1 ,2 ]
机构
[1] Kunming Univ Sci & Technol, Fac Informat Engn & Automat, Kunming 650500, Yunnan, Peoples R China
[2] Yunnan Key Lab Artificial Intelligence, Kunming 650500, Yunnan, Peoples R China
基金
中国国家自然科学基金;
关键词
Event causality identification; Event knowledge; Multilingual pre-trained language models; Parameter-sharing adapter; Pre-training; GRAPH;
D O I
10.1007/s13042-024-02367-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Event causality identification (ECI) aims to identify causal relations between events in texts. Although existing event causality identification works based on fine-tuning pre-trained language models (PLMs) have achieved promising results, they suffer from prohibitive computation costs, catastrophic forgetting of distributional knowledge, as well as poor interpretability. Particularly in low-resource and cross-linguistic scenarios, existing multi-lingual models are generally confronted with the so-called curse of multilinguality, language bias, and hence result in low accuracy and generalization ability. In this paper, we propose a paradigm, termed Pre-training with Event Knowledge of ConceptNet (PTEKC), to couple Multi-lingual Pre-trained Language Models (mPLMs) with event knowledge for cross-lingual event causality identification. Specifically, we have develop a parameter-sharing adapter plugin that facilitates the integration of event knowledge into the frozen PLMs. This approach significantly diminishes the number of trainable parameters and greatly reduces the risk of catastrophic forgetting. Our Adapter integrates multi-lingual alignment event knowledge into the mPLMs through two designed pre-training tasks, namely event masking and self-supervised link prediction. Extensive experiments on the benchmark dataset MECI show that PTEKC is parameter-efficient and can effectively incorporate multi-lingual alignment event knowledge for improving cross-lingual event causality identification.
引用
收藏
页码:1859 / 1872
页数:14
相关论文
共 50 条
  • [21] XLM-E: Cross-lingual Language Model Pre-training via ELECTRA
    Chi, Zewen
    Huang, Shaohan
    Dong, Li
    Ma, Shuming
    Zheng, Bo
    Singhal, Saksham
    Bajaj, Payal
    Song, Xia
    Mao, Xian-Ling
    Huang, Heyan
    Wei, Furu
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 6170 - 6182
  • [22] INFOXLM: An Information-Theoretic Framework for Cross-Lingual Language Model Pre-Training
    Chi, Zewen
    Dong, Li
    Wei, Furu
    Yang, Nan
    Singhal, Saksham
    Wang, Wenhui
    Song, Xia
    Mao, Xian-Ling
    Huang, Heyan
    Zhou, Ming
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 3576 - 3588
  • [23] Event Camera Data Pre-training
    Yang, Yan
    Pan, Liyuan
    Liu, Liu
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 10665 - 10675
  • [24] Integrating Event Elements for Chinese-Vietnamese Cross-Lingual Event Retrieval
    Huang, Yuxin
    Yang, Yuanlin
    Zhu, Enchang
    Liang, Yin
    Xian, Yantuan
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2024, E107D (10) : 1353 - 1361
  • [25] Exploiting Cross-Lingual Hints to Discover Event Pronouns
    Loaiciga, Sharid
    Hardmeier, Christian
    Sayeed, Asad
    PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2020), 2020, : 99 - 103
  • [26] Cross-lingual Structure Transfer for Relation and Event Extraction
    Subburathinam, Ananya
    Lu, Di
    Ji, Heng
    May, Jonathan
    Chang, Shih-Fu
    Sil, Avirup
    Voss, Clare
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 313 - 325
  • [27] Language Model Priming for Cross-Lingual Event Extraction
    Fincke, Steven
    Agarwal, Shantanu
    Miller, Scott
    Boschee, Elizabeth
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 10627 - 10635
  • [28] Cross-View Language Modeling: Towards Unified Cross-Lingual Cross-Modal Pre-training
    Zeng, Yan
    Zhou, Wangchunshu
    Luo, Ao
    Cheng, Ziming
    Zhang, Xinsong
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 5731 - 5746
  • [29] Few-Shot Cross-Lingual Stance Detection with Sentiment-Based Pre-training
    Hardalov, Momchil
    Arora, Arnav
    Nakov, Preslav
    Augenstein, Isabelle
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 10729 - 10737
  • [30] Chinese-Vietnamese cross-lingual event retrieval method based on knowledge distillation
    Gao S.
    He Z.
    Yu Z.
    Zhu E.
    Wu S.
    Journal of Intelligent and Fuzzy Systems, 2024, 46 (04): : 8461 - 8475