Cross-sentence Pre-trained model for Interactive QA matching

被引:0
|
作者
Wu, Jinmeng [1 ,2 ]
Hao, Yanbin [3 ]
机构
[1] Wuhan Inst Technol, Sch Elect & Informat Engn, Wuhan, Peoples R China
[2] Univ Liverpool, Sch Elect Engn Elect & Comp Sci, Brownlow Hill, Liverpool, Merseyside, England
[3] City Univ Hong Kong, Dept Comp Sci, Hong Kong 999077, Peoples R China
基金
中国国家自然科学基金;
关键词
Question answering; Interactive matching; Pre-trained language model; Context jump Dependencies;
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Semantic matching measures the dependencies between query and answer representations, which is an important criterion for evaluating whether the matching is successful. In fact, such matching does not examine each sentence individually, because the context information between sentences should be considered equally important to the syntactic context inside a sentence. Considering the above, we propose a novel QA matching model, built upon a cross-sentence context-aware architecture. Specifically, an interactive attention mechanism with a pre-trained language model is presented to automatically select salient positional answer representations that contribute more significantly to the answer relevance of a given question. In addition to the context information captured at each word position, we incorporate a quantity of context jump dependencies to leverage the attention weight formulation. This can capture the amount of useful information brought by the next word, is computed by modeling the joint probability between two adjacent word states. The proposed method is compared with multiple state-of-the-art methods using the TREC library, WikiQA, and the Yahoo! community question datasets. Experimental results show that the proposed method outperforms satisfactorily the competing ones.
引用
收藏
页码:5417 / 5424
页数:8
相关论文
共 50 条
  • [41] PTMA: Pre-trained Model Adaptation for Transfer Learning
    Li, Xiao
    Yan, Junkai
    Jiang, Jianjian
    Zheng, Wei-Shi
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT I, KSEM 2024, 2024, 14884 : 176 - 188
  • [42] Pre-trained Language Model for Biomedical Question Answering
    Yoon, Wonjin
    Lee, Jinhyuk
    Kim, Donghyeon
    Jeong, Minbyul
    Kang, Jaewoo
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2019, PT II, 2020, 1168 : 727 - 740
  • [43] BERTweet: A pre-trained language model for English Tweets
    Dat Quoc Nguyen
    Thanh Vu
    Anh Tuan Nguyen
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING: SYSTEM DEMONSTRATIONS, 2020, : 9 - 14
  • [44] Tuning Pre-trained Model via Moment Probing
    Gao, Mingze
    Wang, Qilong
    Lin, Zhenyi
    Zhu, Pengfei
    Hu, Qinghua
    Zhou, Jingbo
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 11769 - 11779
  • [45] Solving ESL Sentence Completion Questions via Pre-trained Neural Language Models
    Liu, Qiongqiong
    Liu, Tianqiao
    Zhao, Jiafu
    Fang, Qiang
    Ding, Wenbiao
    Wu, Zhongqin
    Xia, Feng
    Tang, Jiliang
    Liu, Zitao
    ARTIFICIAL INTELLIGENCE IN EDUCATION (AIED 2021), PT II, 2021, 12749 : 256 - 261
  • [46] Session Search with Pre-trained Graph Classification Model
    Ma, Shengjie
    Chen, Chong
    Mao, Jiaxin
    Tian, Qi
    Jiang, Xuhui
    PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, : 953 - 962
  • [47] A Dynamic pre-trained Model for Chinese Classical Poetry
    Wang, Xiaotong
    Liu, Xuanning
    Wang, Haorui
    Wu, Bin
    DATABASE SYSTEMS FOR ADVANCED APPLICATIONS, DASFAA 2024, PT 2, 2025, 14851 : 436 - 445
  • [48] OMPGPT: A Generative Pre-trained Transformer Model for OpenMP
    Chen, Le
    Bhattacharjee, Arijit
    Ahmed, Nesreen
    Hasabnis, Niranjan
    Oren, Gal
    Vo, Vy
    Jannesari, Ali
    EURO-PAR 2024: PARALLEL PROCESSING, PT I, EURO-PAR 2024, 2024, 14801 : 121 - 134
  • [49] Online Active Model Selection for Pre-trained Classifiers
    Karimi, Mohammad Reza
    Guerel, Nezihe Merve
    Karlas, Bojan
    Rausch, Johannes
    Zhang, Ce
    Krause, Andreas
    24TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS (AISTATS), 2021, 130 : 307 - +
  • [50] Misspelling Correction with Pre-trained Contextual Language Model
    Hu, Yifei
    Ting, Xiaonan
    Ko, Youlim
    Rayz, Julia Taylor
    PROCEEDINGS OF 2020 IEEE 19TH INTERNATIONAL CONFERENCE ON COGNITIVE INFORMATICS & COGNITIVE COMPUTING (ICCI*CC 2020), 2020, : 144 - 149