Enhanced distance-aware self-attention and multi-level match for sentence semantic matching

被引:6
|
作者
Deng, Yao [1 ]
Li, Xianfeng [1 ]
Zhang, Mengyan [1 ]
Lu, Xin [1 ]
Sun, Xia [1 ]
机构
[1] Northwest Univ, Sch Informat Sci & Technol, Xian 710127, Peoples R China
关键词
Sentence Matching; Distance-aware Self-attention; Multi-level Match; Representation learning; NETWORK;
D O I
10.1016/j.neucom.2022.05.103
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Sentence semantic matching is a core research area in natural language processing, which is widely used in various natural language tasks. In recent years, attention mechanism has shown good performance in deep neural networks for sentence semantic matching. Most of the attention-based deep neural networks focus on sentences interaction which ignore modeling the core semantic of the sentence. In other words, they do not consider the importance of the relative distance of words when modeling the sentence semantics, which leads to deviations in modeling the core semantics of the sentence and unstable sentence interaction. Usually, people tend to associate words that are relatively close together when they read and believe that there is a deeper connection between them. Besides, the current interactive matching method after sentence modeling is relatively simple and it may be inadequate. In this paper, we build a well-performed distance-aware self-attention and multi-level matching model (DSSTM) for sentence semantic matching tasks. By considering the importance of different distance tokens, it can get the better original semantics of sentences and hold interactive matching method in multiple level after sentence modeling. To be specific, given two input sentences, we first encode them as contextual embeddings. Then, the contextual embeddings are handled by enhanced distance-aware self-attention to further strengthen the sentence semantic modeling from the whole and local aspect. At the same time, we apply the co-attention layer to extract cross-sentence interaction features while simplifying all the remaining components. Finally, we fuse them into the multi-level matching function to obtain the aggregation vector and learn divers matching representations, which is helpful to capture the diversity of sentence pairs. We conduct experiments on three sentence semantic matching tasks. Experimental results on these public datasets demonstrate that our model outperforms competitive baseline methods and our model has fewer parameters. Our source code is publicly available at https://github.com/xiaodeng-1/DSSTM.(c) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页码:174 / 187
页数:14
相关论文
共 50 条
  • [1] Energy-efficient multi-level and distance-aware clustering mechanism for WSNs
    Mehmood, Amjad
    Khan, S.
    Shams, Bilal
    Lloret, Jaime
    INTERNATIONAL JOURNAL OF COMMUNICATION SYSTEMS, 2015, 28 (05) : 972 - 989
  • [2] Channel2DTransformer: A Multi-level Features Self-attention Fusion Module for Semantic Segmentation
    Liu, Weitao
    Wu, Junjun
    INTERNATIONAL JOURNAL OF COMPUTATIONAL INTELLIGENCE SYSTEMS, 2024, 17 (01)
  • [3] Sentence Matching with Deep Self-attention and Co-attention Features
    Wang, Zhipeng
    Yan, Danfeng
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, KSEM 2021, PT II, 2021, 12816 : 550 - 561
  • [4] SATSal: A Multi-Level Self-Attention Based Architecture for Visual Saliency Prediction
    Tliba, Marouane
    Kerkouri, Mohamed A.
    Ghariba, Bashir
    Chetouani, Aladine
    Coeltekin, Arzu
    Shehata, Mohamed
    Bruno, Alessandro
    IEEE ACCESS, 2022, 10 : 20701 - 20713
  • [5] Retinal blood vessel segmentation and inpainting networks with multi-level self-attention
    Golias, Matus
    Sikudova, Elena
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2025, 102
  • [6] Cascaded feature fusion with multi-level self-attention mechanism for object detection
    Wang, Chuanxu
    Wang, Huiru
    PATTERN RECOGNITION, 2023, 138
  • [7] Self-Attention Enhanced Recurrent Neural Networks for Sentence Classification
    Kumar, Ankit
    Rastogi , Reshma
    2018 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI), 2018, : 905 - 911
  • [8] A Multi-level Attention Model for Text Matching
    Sun, Qiang
    Wu, Yue
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2018, PT I, 2018, 11139 : 142 - 153
  • [9] Mention Distance-aware Interactive Attention with Multi-step Reasoning for document-level relation extraction
    Zhang, Fu
    Wang, Jiapeng
    Xu, Huangming
    Wu, Honglin
    Cheng, Jingwei
    Li, Weijun
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2025, 141
  • [10] Multi-level feature fusion capsule network with self-attention for facial expression recognition
    Huang, Zhiji
    Yu, Songsen
    Liang, Jun
    JOURNAL OF ELECTRONIC IMAGING, 2023, 32 (02)