Natural Questions: A Benchmark for Question Answering Research

被引:0
|
作者
Kwiatkowski T. [1 ]
Palomaki J. [1 ]
Redfield O. [1 ]
Collins M. [1 ,2 ]
Parikh A. [1 ]
Alberti C. [1 ]
Epstein D. [1 ]
Polosukhin I. [1 ]
Devlin J. [1 ]
Lee K. [1 ]
Toutanova K. [1 ]
Jones L. [1 ]
Kelcey M. [1 ]
Chang M.-W. [1 ]
Dai A.M. [1 ]
Uszkoreit J. [1 ]
Le Q. [1 ]
Petrov S. [1 ]
机构
[1] Google Research, United States
[2] Columbia University, United States
关键词
D O I
10.1162/tacl_a_00276
中图分类号
学科分类号
摘要
We present the Natural Questions corpus, a question answering data set. Questions consist of real anonymized, aggregated queries issued to the Google search engine. An annotator is presented with a question along with a Wikipedia page from the top 5 search results, and annotates a long answer (typically a paragraph) and a short answer (one or more entities) if present on the page, or marks null if no long/short answer is present. The public release consists of 307,373 training examples with single annotations; 7,830 examples with 5-way annotations for development data; and a further 7,842 examples with 5-way annotated sequestered as test data. We present experiments validating quality of the data. We also describe analysis of 25-way annotations on 302 examples, giving insights into human variability on the annotation task. We introduce robust metrics for the purposes of evaluating question answering systems; demonstrate high human upper bounds on these metrics; and establish baseline results using competitive methods drawn from related literature. © 2019 Association for Computational Linguistics. Distributed under a CC-BY 4.0 license.
引用
收藏
页码:453 / 466
页数:13
相关论文
共 50 条
  • [21] Research on Question Classification for Automatic Question Answering
    Xu, Shihua
    Cheng, Gang
    Kong, Fang
    PROCEEDINGS OF THE 2016 INTERNATIONAL CONFERENCE ON ASIAN LANGUAGE PROCESSING (IALP), 2016, : 218 - 221
  • [22] Questioning the Question - Addressing the Answerability of Questions in Community Question-Answering
    Shah, Chirag
    Kitzie, Vanessa
    Choi, Erik
    2014 47TH HAWAII INTERNATIONAL CONFERENCE ON SYSTEM SCIENCES (HICSS), 2014, : 1386 - 1395
  • [23] Identifying Similar Questions in Healthcare Social Question Answering Services: A Design Science Research Completed Research
    John, Blooma
    Wickramasinghe, Nilmini
    Kurian, Jayan
    AMCIS 2018 PROCEEDINGS, 2018,
  • [24] Developing a Knowledge Graph for a Question Answering System to Answer Natural Language Questions on German Grammar
    Falke, Stefan
    SEMANTIC WEB: ESWC 2019 SATELLITE EVENTS, 2019, 11762 : 199 - 208
  • [25] Question #72. Answering kids' science questions
    Neuenschwander, D. E.
    American Journal of Physics, 66 (04):
  • [26] Personalized Recommendation for New Questions in Community Question Answering
    Wang, Lin
    Wu, Bin
    Yang, Juan
    Peng, Shuang
    PROCEEDINGS OF THE 2016 IEEE/ACM INTERNATIONAL CONFERENCE ON ADVANCES IN SOCIAL NETWORKS ANALYSIS AND MINING ASONAM 2016, 2016, : 901 - 908
  • [27] Named entity disambiguation for questions in community question answering
    Wang, Fang
    Wu, Wei
    Li, Zhoujun
    Zhou, Ming
    KNOWLEDGE-BASED SYSTEMS, 2017, 126 : 68 - 77
  • [28] Improving the clarity of questions in Community Question Answering networks
    Khabbazan, Alireza
    Abin, Ahmad Ali
    Vu, Viet-Vu
    JOURNAL OF INTELLIGENT INFORMATION SYSTEMS, 2024, 62 (6) : 1631 - 1658
  • [29] Recursive Question Decomposition for Answering Complex Geographic Questions
    Hartrumpf, Sven
    Leveling, Johannes
    MULTILINGUAL INFORMATION ACCESS EVALUATION I: TEXT RETRIEVAL EXPERIMENTS, 2010, 6241 : 310 - +
  • [30] Answering contextual questions based on ontologies and question templates
    Dongsheng Wang
    Frontiers of Computer Science in China, 2011, 5 : 405 - 418