Cross-sentence Pre-trained model for Interactive QA matching

被引:0
|
作者
Wu, Jinmeng [1 ,2 ]
Hao, Yanbin [3 ]
机构
[1] Wuhan Inst Technol, Sch Elect & Informat Engn, Wuhan, Peoples R China
[2] Univ Liverpool, Sch Elect Engn Elect & Comp Sci, Brownlow Hill, Liverpool, Merseyside, England
[3] City Univ Hong Kong, Dept Comp Sci, Hong Kong 999077, Peoples R China
基金
中国国家自然科学基金;
关键词
Question answering; Interactive matching; Pre-trained language model; Context jump Dependencies;
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Semantic matching measures the dependencies between query and answer representations, which is an important criterion for evaluating whether the matching is successful. In fact, such matching does not examine each sentence individually, because the context information between sentences should be considered equally important to the syntactic context inside a sentence. Considering the above, we propose a novel QA matching model, built upon a cross-sentence context-aware architecture. Specifically, an interactive attention mechanism with a pre-trained language model is presented to automatically select salient positional answer representations that contribute more significantly to the answer relevance of a given question. In addition to the context information captured at each word position, we incorporate a quantity of context jump dependencies to leverage the attention weight formulation. This can capture the amount of useful information brought by the next word, is computed by modeling the joint probability between two adjacent word states. The proposed method is compared with multiple state-of-the-art methods using the TREC library, WikiQA, and the Yahoo! community question datasets. Experimental results show that the proposed method outperforms satisfactorily the competing ones.
引用
收藏
页码:5417 / 5424
页数:8
相关论文
共 50 条
  • [21] Interactive Design by Integrating a Large Pre-Trained Language Model and Building Information Modeling
    Jang, Suhyung
    Lee, Ghang
    COMPUTING IN CIVIL ENGINEERING 2023-VISUALIZATION, INFORMATION MODELING, AND SIMULATION, 2024, : 291 - 299
  • [22] Pre-Trained Language Models for Interactive Decision-Making
    Li, Shuang
    Puig, Xavier
    Paxton, Chris
    Du, Yilun
    Wang, Clinton
    Fan, Linxi
    Chen, Tao
    Huang, De-An
    Akyurek, Ekin
    Anandkumar, Anima
    Andreas, Jacob
    Mordatch, Igor
    Torralba, Antonio
    Zhu, Yuke
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [23] Talent Supply and Demand Matching Based on Prompt Learning and the Pre-Trained Language Model
    Li, Kunping
    Liu, Jianhua
    Zhuang, Cunbo
    APPLIED SCIENCES-BASEL, 2025, 15 (05):
  • [24] Probing the Robustness of Pre-trained Language Models for Entity Matching
    Rastaghi, Mehdi Akbarian
    Kamalloo, Ehsan
    Rafiei, Davood
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 3786 - 3790
  • [25] Patent classification with pre-trained Bert model
    Kahraman, Selen Yuecesoy
    Durmusoglu, Alptekin
    Dereli, Tuerkay
    JOURNAL OF THE FACULTY OF ENGINEERING AND ARCHITECTURE OF GAZI UNIVERSITY, 2024, 39 (04): : 2485 - 2496
  • [26] Adder Encoder for Pre-trained Language Model
    Ding, Jianbang
    Zhang, Suiyun
    Li, Linlin
    CHINESE COMPUTATIONAL LINGUISTICS, CCL 2023, 2023, 14232 : 339 - 347
  • [27] Harnessing Pre-Trained Sentence Transformers for Offensive Language Detection in Indian Languages
    MKSSS Cummins College of Engineering for Women, Maharashtra, Pune, India
    不详
    不详
    CEUR Workshop Proc., (427-434):
  • [28] Quality Estimation and Translation Metrics via Pre-trained Word and Sentence Embeddings
    Yankovskaya, Elizaveta
    Tattar, Andre
    Fishel, Mark
    FOURTH CONFERENCE ON MACHINE TRANSLATION (WMT 2019), VOL 3: SHARED TASK PAPERS, DAY 2, 2019, : 101 - 105
  • [29] Considering Nested Tree Structure in Sentence Extractive Summarization with Pre-trained Transformer
    Kwon, Jingun
    Kobayashi, Naoki
    Kamigaito, Hidetaka
    Okumura, Manabu
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 4039 - 4044
  • [30] TANDA: Transfer and Adapt Pre-Trained Transformer Models for Answer Sentence Selection
    Garg, Siddhant
    Vu, Thuy
    Moschitti, Alessandro
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 7780 - 7788