DE-ESD: Dual encoder-based entity synonym discovery using pre-trained contextual embeddings

被引:0
|
作者
Huang, Subin [1 ]
Chen, Junjie [1 ]
Yu, Chengzhen [1 ]
Li, Daoyu [1 ]
Zhou, Qing [1 ]
Liu, Sanmin [1 ]
机构
[1] AnHui Polytech Univ, Sch Comp & Informat, Wuhu 241000, Anhui, Peoples R China
关键词
Entity synonym set; Entity synonym discovery; Dual encoder; Pre-trained language model; Contextual embedding;
D O I
10.1016/j.eswa.2025.127102
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Extracting synonymous entities from unstructured text is important for enhancing entity-dependent applications such as web searches and question-answering systems. Existing work primarily falls into two types: statistics- based and deep learning-based. However, these approaches often fail to discern fine semantic nuances among entity mentions and are prone to cumulative errors; thus, they inadequately represent the holistic semantics of entity synonym sets. To address these limitations, this paper introduces a novel framework, Dual Encoder-based Entity Synonym Discovery (DE-ESD). The proposed method initially uses pre-trained language models to extract multiperspective contextual embeddings of entity mentions. Then, it employs a dual encoder architecture to differentiate features between an established entity synonym set and a pseudo-set-created by adding a candidate entity mention to the synonym set. A set scorer evaluates the quality scores of both sets. By leveraging the trained dual encoder and the set scorer, DE-ESD can implement an efficient online algorithm for mining new entity synonym sets for open text streams. The experimental results obtained on two real-world datasets (NYT and Wiki) demonstrate the effectiveness of DE-ESD. Furthermore, we investigated the impact of different pre-trained language models on DE-ESD performance, particularly their ability to extract effective contextual embeddings.
引用
收藏
页数:18
相关论文
共 4 条
  • [1] A graph-based blocking approach for entity matching using pre-trained contextual embedding models*
    Mugeni, John Bosco
    Amagasa, Toshiyuki
    37TH ANNUAL ACM SYMPOSIUM ON APPLIED COMPUTING, 2022, : 357 - 364
  • [2] Using Pre-trained Deeply Contextual Model BERT for Russian Named Entity Recognition
    Mukhin, Eugeny
    ANALYSIS OF IMAGES, SOCIAL NETWORKS AND TEXTS (AIST 2019), 2020, 1086 : 167 - 173
  • [3] Alzheimer Disease Recognition Using Speech-Based Embeddings From Pre-Trained Models
    Gauder, Lara
    Pepino, Leonardo
    Ferrer, Luciana
    Riera, Pablo
    INTERSPEECH 2021, 2021, : 3795 - 3799
  • [4] DeepAIP: Deep learning for anti-inflammatory peptide prediction using pre-trained protein language model features based on contextual self-attention network
    Zhu, Lun
    Yang, Qingguo
    Yang, Sen
    INTERNATIONAL JOURNAL OF BIOLOGICAL MACROMOLECULES, 2024, 280