Enhancing yes/no question answering with weak supervision via extractive question answering

被引:1
|
作者
Dimitriadis, Dimitris [1 ]
Tsoumakas, Grigorios [1 ]
机构
[1] Aristotle Univ Thessaloniki, Sch Informat, Thessaloniki 54124, Greece
关键词
Question answering; Yes/no question answering; Extractive question answering; Transformers;
D O I
10.1007/s10489-023-04751-w
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The effectiveness of natural language processing models relies on various factors, including the architecture, number of parameters, data used during training, and the tasks they were trained on. Recent studies indicate that models pre-trained on large corpora and fine-tuned on task-specific datasets, covering multiple tasks, can generate remarkable results across various benchmarks. We propose a new approach based on a straightforward hypothesis: improving model performance on a target task by considering other artificial tasks defined on the same training dataset. By doing so, the model can gain further insights into the training dataset and attain a greater understanding, improving efficiency on the target task. This approach differs from others that consider multiple pre-existing tasks on different datasets. We validate this hypothesis by focusing on the problem of answering yes/no questions and introducing a multi-task model that outputs a span of the reference text, serving as evidence for answering the question. The task of span extraction is an artificial one, designed to benefit the performance of the model answering yes/no questions. We acquire weak supervision for these spans, by using a pre-trained extractive question answering model, dispensing the need for costly human annotation. Our experiments, using modern transformer-based language models, demonstrate that this method outperforms the standard approach of training models to answer yes/no questions. Although the primary objective was to enhance the performance of the model in answering yes/no questions, it was discovered that span texts are a significant source of information. These spans, derived from the question reference texts, provided valuable insights for the users to better comprehend the answers to the questions. The model's improved accuracy in answering yes/no questions, coupled with the supplementary information provided by the span texts, led to a more comprehensive and informative user experience.
引用
收藏
页码:27560 / 27570
页数:11
相关论文
共 50 条
  • [1] Enhancing yes/no question answering with weak supervision via extractive question answering
    Dimitris Dimitriadis
    Grigorios Tsoumakas
    Applied Intelligence, 2023, 53 : 27560 - 27570
  • [2] WeaQA: Weak Supervision via Captions for Visual Question Answering
    Banerjee, Pratyay
    Gokhale, Tejas
    Yang, Yezhou
    Baral, Chitta
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 3420 - 3435
  • [3] INTERPRETABLE VISUAL QUESTION ANSWERING VIA REASONING SUPERVISION
    Parelli, Maria
    Mallis, Dimitrios
    Diomataris, Markos
    Pitsikalis, Vassilis
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 2525 - 2529
  • [4] Yes/No Question Answering in BioASQ 2019
    Dimitriadis, Dimitris
    Tsoumakas, Grigorios
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2019, PT II, 2020, 1168 : 661 - 669
  • [5] Factoid Question Answering with Distant Supervision
    Zhang, Hongzhi
    Liang, Xiao
    Xu, Guangluan
    Fu, Kun
    Li, Feng
    Huang, Tinglei
    ENTROPY, 2018, 20 (06)
  • [6] Sequence tagging for biomedical extractive question answering
    Yoon, Wonjin
    Jackson, Richard
    Lagerberg, Aron
    Kang, Jaewoo
    BIOINFORMATICS, 2022, 38 (15) : 3794 - 3801
  • [7] QUASER: Question Answering with Scalable Extractive Rationalization
    Ghoshal, Asish
    Iyer, Srinivasan
    Paranjape, Bhargavi
    Lakhotia, Kushal
    Yih, Scott Wen-tau
    Mehdad, Yashar
    PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, : 1208 - 1218
  • [9] Neural Ranking with Weak Supervision for Open-Domain Question Answering : A Survey
    Shen, Xiaoyu
    Vakulenko, Svitlana
    del Tredici, Marco
    Barlacchi, Gianni
    Byrne, Bill
    de Gispert, Adria
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 1736 - 1750
  • [10] Extractive Question Answering with Contrastive Puzzles and Reweighted Clues
    Liu, Chao
    Yang, Jie
    Li, Wanqing
    DOCUMENT ANALYSIS AND RECOGNITION-ICDAR 2024, PT VI, 2024, 14809 : 97 - 112