Enhanced distance-aware self-attention and multi-level match for sentence semantic matching

被引:6
|
作者
Deng, Yao [1 ]
Li, Xianfeng [1 ]
Zhang, Mengyan [1 ]
Lu, Xin [1 ]
Sun, Xia [1 ]
机构
[1] Northwest Univ, Sch Informat Sci & Technol, Xian 710127, Peoples R China
关键词
Sentence Matching; Distance-aware Self-attention; Multi-level Match; Representation learning; NETWORK;
D O I
10.1016/j.neucom.2022.05.103
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Sentence semantic matching is a core research area in natural language processing, which is widely used in various natural language tasks. In recent years, attention mechanism has shown good performance in deep neural networks for sentence semantic matching. Most of the attention-based deep neural networks focus on sentences interaction which ignore modeling the core semantic of the sentence. In other words, they do not consider the importance of the relative distance of words when modeling the sentence semantics, which leads to deviations in modeling the core semantics of the sentence and unstable sentence interaction. Usually, people tend to associate words that are relatively close together when they read and believe that there is a deeper connection between them. Besides, the current interactive matching method after sentence modeling is relatively simple and it may be inadequate. In this paper, we build a well-performed distance-aware self-attention and multi-level matching model (DSSTM) for sentence semantic matching tasks. By considering the importance of different distance tokens, it can get the better original semantics of sentences and hold interactive matching method in multiple level after sentence modeling. To be specific, given two input sentences, we first encode them as contextual embeddings. Then, the contextual embeddings are handled by enhanced distance-aware self-attention to further strengthen the sentence semantic modeling from the whole and local aspect. At the same time, we apply the co-attention layer to extract cross-sentence interaction features while simplifying all the remaining components. Finally, we fuse them into the multi-level matching function to obtain the aggregation vector and learn divers matching representations, which is helpful to capture the diversity of sentence pairs. We conduct experiments on three sentence semantic matching tasks. Experimental results on these public datasets demonstrate that our model outperforms competitive baseline methods and our model has fewer parameters. Our source code is publicly available at https://github.com/xiaodeng-1/DSSTM.(c) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页码:174 / 187
页数:14
相关论文
共 50 条
  • [41] MuLAN: Multi-level attention-enhanced matching network for few-shot knowledge graph completion
    Li, Qianyu
    Feng, Bozheng
    Tang, Xiaoli
    Yu, Han
    Song, Hengjie
    NEURAL NETWORKS, 2024, 174
  • [42] Multi-level authorisation model and framework for distributed semantic-aware environments
    Amini, M.
    Jalili, R.
    IET INFORMATION SECURITY, 2010, 4 (04) : 301 - 321
  • [43] Multi-level Symmetric Semantic Alignment Network for image-text matching
    Wang, Wenzhuang
    Di, Xiaoguang
    Liu, Maozhen
    Gao, Feng
    NEUROCOMPUTING, 2024, 599
  • [44] Image-text matching algorithm based on multi-level semantic alignment
    Li Y.
    Yao T.
    Zhang L.
    Sun Y.
    Fu H.
    Beijing Hangkong Hangtian Daxue Xuebao/Journal of Beijing University of Aeronautics and Astronautics, 2024, 50 (02): : 551 - 558
  • [45] Domain-Aware Self-Attention for Multi-Domain Neural Machine Translation
    Zhang, Shiqi
    Liu, Yan
    Xiong, Deyi
    Zhang, Pei
    Chen, Boxing
    INTERSPEECH 2021, 2021, : 2047 - 2051
  • [46] Spatial Context-Aware Self-Attention Model For Multi-Organ Segmentation
    Tang, Hao
    Liu, Xingwei
    Han, Kun
    Xie, Xiaohui
    Chen, Xuming
    Qian, Huang
    Liu, Yong
    Sun, Shanlin
    Bai, Narisu
    2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2021), 2021, : 938 - 948
  • [47] A Sparse Self-Attention Enhanced Model for Aspect-Level Sentiment Classification
    Dhanith, P. R. Joe
    Surendiran, B.
    Rohith, G.
    Kanmani, Sujithra R.
    Devi, K. Valli
    NEURAL PROCESSING LETTERS, 2024, 56 (02)
  • [48] A Sparse Self-Attention Enhanced Model for Aspect-Level Sentiment Classification
    P. R. Joe Dhanith
    B. Surendiran
    G. Rohith
    Sujithra R. Kanmani
    K. Valli Devi
    Neural Processing Letters, 56
  • [49] Exploring Multi-Level Attention and Semantic Relationship for Remote Sensing Image Captioning
    Yuan, Zhenghang
    Li, Xuelong
    Wang, Qi
    IEEE ACCESS, 2020, 8 (08): : 2608 - 2620
  • [50] Sentence modeling via multiple word embeddings and multi-level comparison for semantic textual similarity
    Nguyen Huy Tien
    Nguyen Minh Le
    Tomohiro, Yamasaki
    Tatsuya, Izuha
    INFORMATION PROCESSING & MANAGEMENT, 2019, 56 (06)