Cross-sentence Pre-trained model for Interactive QA matching

被引:0
|
作者
Wu, Jinmeng [1 ,2 ]
Hao, Yanbin [3 ]
机构
[1] Wuhan Inst Technol, Sch Elect & Informat Engn, Wuhan, Peoples R China
[2] Univ Liverpool, Sch Elect Engn Elect & Comp Sci, Brownlow Hill, Liverpool, Merseyside, England
[3] City Univ Hong Kong, Dept Comp Sci, Hong Kong 999077, Peoples R China
基金
中国国家自然科学基金;
关键词
Question answering; Interactive matching; Pre-trained language model; Context jump Dependencies;
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Semantic matching measures the dependencies between query and answer representations, which is an important criterion for evaluating whether the matching is successful. In fact, such matching does not examine each sentence individually, because the context information between sentences should be considered equally important to the syntactic context inside a sentence. Considering the above, we propose a novel QA matching model, built upon a cross-sentence context-aware architecture. Specifically, an interactive attention mechanism with a pre-trained language model is presented to automatically select salient positional answer representations that contribute more significantly to the answer relevance of a given question. In addition to the context information captured at each word position, we incorporate a quantity of context jump dependencies to leverage the attention weight formulation. This can capture the amount of useful information brought by the next word, is computed by modeling the joint probability between two adjacent word states. The proposed method is compared with multiple state-of-the-art methods using the TREC library, WikiQA, and the Yahoo! community question datasets. Experimental results show that the proposed method outperforms satisfactorily the competing ones.
引用
收藏
页码:5417 / 5424
页数:8
相关论文
共 50 条
  • [1] Using Pre-trained Language Model to Enhance Active Learning for Sentence Matching
    Bai, Guirong
    He, Shizhu
    Liu, Kang
    Zhao, Jun
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2022, 21 (02)
  • [2] Improving the Performance of Answers Ranking in QA Communities: A Long-Text Matching Technique and Pre-Trained Model
    Sun, Siyu
    Wang, Yiming
    Cheng, Jiale
    Xiao, Zhiying
    Zheng, Daqing
    Hao, Xiaoling
    IEEE ACCESS, 2025, 13 : 4188 - 4200
  • [3] SiBert: Enhanced Chinese Pre-trained Language Model with Sentence Insertion
    Chen, Jiahao
    Cao, Chenjie
    Jiang, Xiuyan
    PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2020), 2020, : 2405 - 2412
  • [4] Interpretability of Entity Matching Based on Pre-trained Language Model
    Liang Z.
    Wang H.-Z.
    Dai J.-J.
    Shao X.-Y.
    Ding X.-O.
    Mu T.-Y.
    Ruan Jian Xue Bao/Journal of Software, 2023, 34 (03): : 1087 - 1108
  • [5] A Cross-Sentence Latent Variable Model for Semi-Supervised Text Sequence Matching
    Choi, Jihun
    Kim, Taeuk
    Lee, Sang-goo
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 4747 - 4761
  • [6] Improving the Performance of Pre-trained Systems in Sentence Retrieval
    Rughbeer, Yastil
    Pillay, Anban W.
    Jembere, Edgar
    2021 IEEE 24TH INTERNATIONAL CONFERENCE ON INFORMATION FUSION (FUSION), 2021, : 908 - 915
  • [7] On the Sentence Embeddings from Pre-trained Language Models
    Li, Bohan
    Zhou, Hao
    He, Junxian
    Wang, Mingxuan
    Yang, Yiming
    Li, Lei
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 9119 - 9130
  • [8] MULTILINGUAL TEXT CLASSIFIER USING PRE-TRAINED UNIVERSAL SENTENCE ENCODER MODEL
    Orlovskiy, O., V
    Sohrab, Khalili
    Ostapov, S. E.
    Hazdyuk, K. P.
    Shumylyak, L. M.
    RADIO ELECTRONICS COMPUTER SCIENCE CONTROL, 2022, (03) : 102 - 108
  • [9] Vietnamese Sentence Paraphrase Identification using Pre-trained Model and Linguistic Knowledge
    Dien Dinh
    Nguyen Le Thanh
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2021, 12 (08) : 796 - 806
  • [10] Vision Enhanced Generative Pre-trained Language Model for Multimodal Sentence Summarization
    Jing, Liqiang
    Li, Yiren
    Xu, Junhao
    Yu, Yongcan
    Shen, Pei
    Song, Xuemeng
    MACHINE INTELLIGENCE RESEARCH, 2023, 20 (02) : 289 - 298