Enhanced distance-aware self-attention and multi-level match for sentence semantic matching

被引:6
|
作者
Deng, Yao [1 ]
Li, Xianfeng [1 ]
Zhang, Mengyan [1 ]
Lu, Xin [1 ]
Sun, Xia [1 ]
机构
[1] Northwest Univ, Sch Informat Sci & Technol, Xian 710127, Peoples R China
关键词
Sentence Matching; Distance-aware Self-attention; Multi-level Match; Representation learning; NETWORK;
D O I
10.1016/j.neucom.2022.05.103
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Sentence semantic matching is a core research area in natural language processing, which is widely used in various natural language tasks. In recent years, attention mechanism has shown good performance in deep neural networks for sentence semantic matching. Most of the attention-based deep neural networks focus on sentences interaction which ignore modeling the core semantic of the sentence. In other words, they do not consider the importance of the relative distance of words when modeling the sentence semantics, which leads to deviations in modeling the core semantics of the sentence and unstable sentence interaction. Usually, people tend to associate words that are relatively close together when they read and believe that there is a deeper connection between them. Besides, the current interactive matching method after sentence modeling is relatively simple and it may be inadequate. In this paper, we build a well-performed distance-aware self-attention and multi-level matching model (DSSTM) for sentence semantic matching tasks. By considering the importance of different distance tokens, it can get the better original semantics of sentences and hold interactive matching method in multiple level after sentence modeling. To be specific, given two input sentences, we first encode them as contextual embeddings. Then, the contextual embeddings are handled by enhanced distance-aware self-attention to further strengthen the sentence semantic modeling from the whole and local aspect. At the same time, we apply the co-attention layer to extract cross-sentence interaction features while simplifying all the remaining components. Finally, we fuse them into the multi-level matching function to obtain the aggregation vector and learn divers matching representations, which is helpful to capture the diversity of sentence pairs. We conduct experiments on three sentence semantic matching tasks. Experimental results on these public datasets demonstrate that our model outperforms competitive baseline methods and our model has fewer parameters. Our source code is publicly available at https://github.com/xiaodeng-1/DSSTM.(c) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页码:174 / 187
页数:14
相关论文
共 50 条
  • [31] Part Matching with Multi-level Attention for Person Re-Identification
    Wang, Jiaze
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW), 2019, : 1805 - 1814
  • [32] MHSAN: Multi-Head Self-Attention Network for Visual Semantic Embedding
    Park, Geondo
    Han, Chihye
    Kim, Daeshik
    Yoon, Wonjun
    2020 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2020, : 1507 - 1515
  • [33] Integration of multi-level semantics in PTMs with an attention model for question matching
    Ye, Zheng
    Che, Linwei
    Ge, Jun
    Qin, Jun
    Liu, Jing
    PLOS ONE, 2024, 19 (08):
  • [34] Contextual-Semantic-Aware Linkable Knowledge Prediction in Stack Overflow via Self-Attention
    Luo, Zhaolin
    Xu, Ling
    Xu, Zhou
    Yan, Meng
    Lei, Yan
    Li, Can
    2021 IEEE 32ND INTERNATIONAL SYMPOSIUM ON SOFTWARE RELIABILITY ENGINEERING (ISSRE 2021), 2021, : 115 - 126
  • [35] Multi-head enhanced self-attention network for novelty detection
    Zhang, Yingying
    Gong, Yuxin
    Zhu, Haogang
    Bai, Xiao
    Tang, Wenzhong
    PATTERN RECOGNITION, 2020, 107
  • [36] SEMScene: Semantic-Consistency Enhanced Multi-Level Scene Graph Matching for Image-Text Retrieval
    Liu, Yuankun
    Yuan, Xiang
    Li, Haochen
    Tan, Zhijie
    Huang, Jinsong
    Xiao, Jingjie
    Li, Weiping
    Mo, Tong
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2024, 20 (08)
  • [37] Multi-Level Visual-Semantic Alignments with Relation-Wise Dual Attention Network for Image and Text Matching
    Hu, Zhibin
    Luo, Yongsheng
    Lin, Jiong
    Yan, Yan
    Chen, Jian
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 789 - 795
  • [38] Multi-Level Head-Wise Match and Aggregation in Transformer for Textual Sequence Matching
    Wang, Shuohang
    Lan, Yunshi
    Tay, Yi
    Jiang, Jing
    Liu, Jingjing
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 9209 - 9216
  • [39] Multi-Level Federated Graph Learning and Self-Attention Based Personalized Wi-Fi Indoor Fingerprint Localization
    Wu, Zheshun
    Wu, Xiaoping
    Long, Yunliang
    IEEE COMMUNICATIONS LETTERS, 2022, 26 (08) : 1794 - 1798
  • [40] MR-FPN: Multi-Level Residual Feature Pyramid Text Detection Network Based on Self-Attention Environment
    Kang, Jianjun
    Ibrayim, Mayire
    Hamdulla, Askar
    SENSORS, 2022, 22 (09)