Enhance prototypical networks with hybrid attention and confusing loss function for few-shot relation classification

被引:10
|
作者
Li, Yibing [1 ,2 ,3 ]
Ma, Zuchang [1 ]
Gao, Lisheng [1 ]
Wu, Yichen [1 ,2 ,4 ]
Xie, Fei [3 ]
Ren, Xiaoye [3 ]
机构
[1] Chinese Acad Sci, Hefei Inst Phys Sci, Inst Intelligent Machines, Anhui Prov Key Lab Med Phys & Technol, Hefei 230031, Peoples R China
[2] Univ Sci & Technol China, Sci Isl Branch Grad Sch, Hefei 230026, Peoples R China
[3] Hefei Normal Univ, Sch Comp Sci & Technol, Hefei 230601, Peoples R China
[4] Anhui Jianzhu Univ, Sch Elect & Informat Engn, Hefei 230601, Peoples R China
基金
中国国家自然科学基金;
关键词
Relation classification; Few-shot learning; Hybrid attention; Loss; BERT;
D O I
10.1016/j.neucom.2022.04.067
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Relation classification (RC) is a fundamental task to building knowledge graphs and describing semantic formalization. It aims to classify a relation between the head and the tail entities in a sentence. The existing RC method mainly adopts the distant supervision (DS) scheme. However, DS still has the problem of long-tail and suffers from data sparsity. Recently, few-shot learning (FSL) has attracted people's attention. It solves the long-tail problem by learning from few-shot samples. The prototypical networks have a better effect on FSL, which classifies a relation by distance. However, the prototypical networks and their related variants did not consider the critical role of entity words. In addition, not all sentences in support set equally contributed to classifying relations. Furthermore, an entity pair in a sentence may have true and confusing relations, which is difficult for the RC model to distinguish them. A new context encoder BERT_FE is proposed to address those problems, which uses the BERT model as pre-training and fuses the information of head and tail entities by entity word-level attention (WLA). At the same time, the sentence-level attention (SLA) is proposed to give more weight to sentences of the support set similar to the query instance and improve the classification accuracy. A confusing loss function (CLF) is designed to enhance the model's ability to distinguish between true and confusing relations. The experiment results demonstrate that our proposed model (HACLF) is better than several baseline models. (c) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页码:362 / 372
页数:11
相关论文
共 50 条
  • [31] Dynamic matching-prototypical learning for noisy few-shot relation classification
    Bi, Haijia
    Peng, Tao
    Han, Jiayu
    Cui, Hai
    Liu, Lu
    KNOWLEDGE-BASED SYSTEMS, 2025, 309
  • [32] Total Relation Network with Attention for Few-Shot Image Classification
    Li X.-X.
    Liu Z.-Y.
    Wu J.-J.
    Cao J.
    Ma Z.-Y.
    Jisuanji Xuebao/Chinese Journal of Computers, 2023, 46 (02): : 371 - 384
  • [33] Enhanced prototypical network for few-shot relation extraction
    Wen, Wen
    Liu, Yongbin
    Ouyang, Chunping
    Lin, Qiang
    Chung, Tonglee
    INFORMATION PROCESSING & MANAGEMENT, 2021, 58 (04)
  • [34] Few-shot classification using Gaussianisation prototypical classifier
    Liu, Fan
    Li, Feifan
    Yang, Sai
    IET COMPUTER VISION, 2023, 17 (01) : 62 - 75
  • [35] ProtoMed: Prototypical networks with auxiliary regularization for few-shot medical image classification
    Ouahab, Achraf
    Ben Ahmed, Olfa
    IMAGE AND VISION COMPUTING, 2025, 154
  • [36] Graph Prototypical Networks for Few-shot Learning on Attributed Networks
    Ding, Kaize
    Wang, Jianling
    Li, Jundong
    Shu, Kai
    Liu, Chenghao
    Liu, Huan
    CIKM '20: PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, 2020, : 295 - 304
  • [37] Knowledge-Enhanced Prototypical Network with Structural Semantics for Few-Shot Relation Classification
    Li, Yanhu
    Zhang, Taolin
    Li, Dongyang
    He, Xiaofeng
    ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2023, PT III, 2023, 13937 : 138 - 149
  • [38] Cost-effective CNNs-based prototypical networks for few-shot relation classification across domains
    Yin, Gongzhu
    Wang, Xing
    Zhang, Hongli
    Wang, Jinlin
    KNOWLEDGE-BASED SYSTEMS, 2022, 253
  • [39] DPNet: domain-aware prototypical network for interdisciplinary few-shot relation classification
    Lv, Bo
    Jin, Li
    Li, Xiaoyu
    Sun, Xian
    Guo, Zhi
    Zhang, Zequn
    Li, Shuchao
    APPLIED INTELLIGENCE, 2022, 52 (13) : 15718 - 15733
  • [40] DPNet: domain-aware prototypical network for interdisciplinary few-shot relation classification
    Bo Lv
    Li Jin
    Xiaoyu Li
    Xian Sun
    Zhi Guo
    Zequn Zhang
    Shuchao Li
    Applied Intelligence, 2022, 52 : 15718 - 15733