A Multi-Task Semantic Decomposition Framework with Task-specific Pre-training for Few-Shot NER

被引:7
|
作者
Dong, Guanting [1 ]
Wang, Zechen [1 ]
Zhao, Jinxu [1 ]
Zhao, Gang [1 ]
Guo, Daichi [1 ]
Fu, Dayuan [1 ]
Hui, Tingfeng [1 ]
Zeng, Chen [1 ]
He, Keqing [2 ]
Li, Xuefeng [1 ]
Wang, Liwen [1 ]
Cui, Xinyue [1 ]
Xu, Weiran [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Beijing, Peoples R China
[2] Meituan Grp Beijing, Beijing, Peoples R China
关键词
Few-shot NER; Multi-Task; Semantic Decomposition; Pre-training;
D O I
10.1145/3583780.3614766
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The objective of few-shot named entity recognition is to identify named entities with limited labeled instances. Previous works have primarily focused on optimizing the traditional token-wise classification framework, while neglecting the exploration of information based on NER data characteristics. To address this issue, we propose a Multi-Task Semantic Decomposition Framework via Joint Task-specific Pre-training (MSDP) for few-shot NER. Drawing inspiration from demonstration-based and contrastive learning, we introduce two novel pre-training tasks: Demonstration-based Masked Language Modeling (MLM) and Class Contrastive Discrimination. These tasks effectively incorporate entity boundary information and enhance entity representation in Pre-trained Language Models (PLMs). In the downstream main task, we introduce a multitask joint optimization framework with the semantic decomposing method, which facilitates the model to integrate two different semantic information for entity classification. Experimental results of two few-shot NER benchmarks demonstrate that MSDP consistently outperforms strong baselines by a large margin. Extensive analyses validate the effectiveness and generalization of MSDP.
引用
收藏
页码:430 / 440
页数:11
相关论文
共 50 条
  • [41] A Kernel Approach to Multi-Task Learning with Task-Specific Kernels
    武威
    李航
    胡云华
    金榕
    Journal of Computer Science & Technology, 2012, 27 (06) : 1289 - 1301
  • [42] A Kernel Approach to Multi-Task Learning with Task-Specific Kernels
    Wei Wu
    Hang Li
    Yun-Hua Hu
    Rong Jin
    Journal of Computer Science and Technology, 2012, 27 : 1289 - 1301
  • [43] Projected Task-Specific Layers for Multi-Task Reinforcement Learning
    Roberts, Josselin Somerville
    Di, Julia
    2024 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA 2024, 2024, : 2887 - 2893
  • [44] Improving AMR-to-text Generation with Multi-task Pre-training
    Xu D.-Q.
    Li J.-H.
    Zhu M.-H.
    Zhou G.-D.
    Ruan Jian Xue Bao/Journal of Software, 2021, 32 (10): : 3036 - 3050
  • [45] Multi-Task Pre-Training for Plug-and-Play Task-Oriented Dialogue System
    Su, Yixuan
    Shu, Lei
    Mansimov, Elman
    Gupta, Arshit
    Cai, Deng
    Lai, Yi-An
    Zhang, Yi
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 4661 - 4676
  • [46] Multi-Task Pre-Training of Deep Neural Networks for Digital Pathology
    Mormont, Romain
    Geurts, Pierre
    Maree, Raphael
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2021, 25 (02) : 412 - 421
  • [47] Multi-task Pre-training for Lhasa-Tibetan Speech Recognition
    Liu, Yigang
    Zhao, Yue
    Xu, Xiaona
    Xu, Liang
    Zhang, Xubei
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT IX, 2023, 14262 : 78 - 90
  • [48] Improving News Recommendation via Bottlenecked Multi-task Pre-training
    Xiao, Xiongfeng
    Li, Qing
    Liu, Songlin
    Zhou, Kun
    PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, : 2082 - 2086
  • [49] MULTI-TASK SELF-SUPERVISED PRE-TRAINING FOR MUSIC CLASSIFICATION
    Wu, Ho-Hsiang
    Kao, Chieh-Chi
    Tang, Qingming
    Sun, Ming
    McFee, Brian
    Bello, Juan Pablo
    Wang, Chao
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 556 - 560
  • [50] MVP: Multi-task Supervised Pre-training for Natural Language Generation
    Tang, Tianyi
    Li, Junyi
    Zhao, Wayne Xin
    Wen, Ji-Rong
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 8758 - 8794