QED: A Framework and Dataset for Explanations in Question Answering

被引:14
|
作者
Lamm, Matthew [1 ,4 ]
Palomaki, Jennimaria [2 ]
Alberti, Chris [2 ]
Andor, Daniel [2 ]
Choi, Eunsol [3 ,4 ]
Soares, Livio Baldini [2 ]
Collins, Michael [2 ]
机构
[1] Stanford Univ, Dept Linguist, Stanford, CA 94305 USA
[2] Google Res, Mountain View, CA USA
[3] Univ Texas Austin, Dept Comp Sci, Austin, TX 78712 USA
[4] Google, Mountain View, CA 94043 USA
关键词
D O I
10.1162/tacl_a_00398
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A question answering system that in addition to providing an answer provides an explanation of the reasoning that leads to that answer has potential advantages in terms of debuggability, extensibility, and trust. To this end, we propose QED, a linguistically informed, extensible framework for explanations in question answering. A QED explanation specifies the relationship between a question and answer according to formal semantic notions such as referential equality, sentencehood, and entailment. We describe and publicly release an expert-annotated dataset of QED explanations built upon a subset of the Google Natural Questions dataset, and report baselinemodels on two tasks-post-hoc explanation generation given an answer, and joint question answering and explanation generation. In the joint setting, a promising result suggests that training on a relatively small amount of QED data can improve question answering. In addition to describing the formal, language-theoretic motivations for the QED approach, we describe a large user study showing that the presence of QED explanations significantly improves the ability of untrained raters to spot errors made by a strong neural QA baseline.
引用
收藏
页码:790 / 806
页数:17
相关论文
共 50 条
  • [21] A Portuguese Dataset for Evaluation of Semantic Question Answering
    de Araujo, Denis Andrei
    Rigo, Sandro Jose
    Quaresma, Paulo
    Muniz, Joao Henrique
    COMPUTATIONAL PROCESSING OF THE PORTUGUESE LANGUAGE, PROPOR 2020, 2020, 12037 : 217 - 227
  • [22] Single-dataset Experts for Multi-dataset Question Answering
    Friedman, Dan
    Dodge, Ben
    Chen, Danqi
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 6128 - 6137
  • [23] Dataset bias: A case study for visual question answering
    Das A.
    Anjum S.
    Gurari D.
    Proceedings of the Association for Information Science and Technology, 2019, 56 (01): : 58 - 67
  • [24] Improvisation of Dataset Efficiency in Visual Question Answering Domain
    Mohamed, Sheerin Sitara Noor
    Srinivasan, Kavitha
    STATISTICS AND APPLICATIONS, 2022, 20 (02): : 279 - 289
  • [25] RuBQ 2.0: An Innovated Russian Question Answering Dataset
    Rybin, Ivan
    Korablinov, Vladislav
    Efimov, Pavel
    Braslavski, Pavel
    SEMANTIC WEB, ESWC 2021, 2021, 12731 : 532 - 547
  • [26] Building a benchmark dataset for the Kurdish news question answering
    Saeed, Ari M.
    DATA IN BRIEF, 2024, 57
  • [27] EgoVQA - An Egocentric Video Question Answering Benchmark Dataset
    Fan, Chenyou
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW), 2019, : 4359 - 4366
  • [28] A dataset for medical instructional video classification and question answering
    Gupta, Deepak
    Attal, Kush
    Demner-Fushman, Dina
    SCIENTIFIC DATA, 2023, 10 (01)
  • [29] DermaVQA: A Multilingual Visual Question Answering Dataset for Dermatology
    Yim, Wen-wai
    Fu, Yujuan
    Sun, Zhaoyi
    Ben Abacha, Asma
    Yetisgen, Meliha
    Xia, Fei
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2024, PT V, 2024, 15005 : 209 - 219
  • [30] OVQA: A Clinically Generated Visual Question Answering Dataset
    Huang, Yefan
    Wang, Xiaoli
    Liu, Feiyan
    Huang, Guofeng
    PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, : 2924 - 2938