DISFL-QA: A Benchmark Dataset for Understanding Disfluencies in Question Answering

被引:0
|
作者
Gupta, Aditya [1 ]
Xu, Jiacheng [2 ,4 ]
Upadhyay, Shyam [1 ]
Yang, Diyi [3 ]
Faruqui, Manaal [1 ]
机构
[1] Google Assistant, Mountain View, CA USA
[2] Univ Texas Austin, Austin, TX 78712 USA
[3] Georgia Inst Technol, Atlanta, GA 30332 USA
[4] Google, Mountain View, CA USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Disfluencies is an under-studied topic in NLP, even though it is ubiquitous in human conversation. This is largely due to the lack of datasets containing disfluencies. In this paper, we present a new challenge question answering dataset, DISFL-QA, a derivative of SQUAD, where humans introduce contextual disfluencies in previously fluent questions. DISFL- QA contains a variety of challenging disfluencies that require a more comprehensive understanding of the text than what was necessary in prior datasets. Experiments show that the performance of existing state-of-the-art question answering models degrades significantly when tested on DISFLQA in a zero-shot setting. We show data augmentation methods partially recover the loss in performance and also demonstrate the efficacy of using gold data for fine-tuning. We argue that we need large-scale disfluency datasets in order for NLP models to be robust to them. The dataset is publicly available at: https://github.com/ google-research-datasets/disfl-qa.
引用
收藏
页码:3309 / 3319
页数:11
相关论文
共 50 条
  • [21] Event-QA: A Dataset for Event-Centric Question Answering over Knowledge Graphs
    Costa, Tarcisio Souza
    Gottschalk, Simon
    Demidova, Elena
    CIKM '20: PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, 2020, : 3157 - 3164
  • [22] Automatic question answering for multiple stakeholders, the epidemic question answering dataset
    Travis R. Goodwin
    Dina Demner-Fushman
    Kyle Lo
    Lucy Lu Wang
    Hoa T. Dang
    Ian M. Soboroff
    Scientific Data, 9
  • [23] Automatic question answering for multiple stakeholders, the epidemic question answering dataset
    Goodwin, Travis R.
    Demner-Fushman, Dina
    Lo, Kyle
    Wang, Lucy Lu
    Dang, Hoa T.
    Soboroff, Ian M.
    SCIENTIFIC DATA, 2022, 9 (01)
  • [24] NuScenes-QA: A Multi-Modal Visual Question Answering Benchmark for Autonomous Driving Scenario
    Qian, Tianwen
    Chen, Jingjing
    Zhuo, Linhai
    Jiao, Yang
    Jiang, Yu-Gang
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 5, 2024, : 4542 - 4550
  • [25] (QA)2: Question Answering with Questionable Assumptions
    Kim, Najoung
    Phu Mon Htut
    Bowman, Samuel R.
    Petty, Jackson
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 8466 - 8487
  • [26] TempQuestions: A Benchmark for Temporal Question Answering
    Jia, Zhen
    Abujabal, Abdalghani
    Roy, Rishiraj Saha
    Stroetgen, Jannik
    Weikum, Gerhard
    COMPANION PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE 2018 (WWW 2018), 2018, : 1057 - 1062
  • [27] XMQAs: Constructing Complex-Modified Question-Answering Dataset for Robust Question Understanding
    Chen, Yuyan
    Xiao, Yanghua
    Li, Zhixu
    Liu, Bang
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (03) : 1371 - 1384
  • [28] Question Answering System Based on Community QA
    Zheng, Zhicheng
    Tang, Yang
    Long, Chong
    Bu, Fan
    Zhu, Xiaoyan
    LREC 2010 - SEVENTH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2010, : D23 - D27
  • [29] Reproducing a Neural Question Answering Architecture Applied to the SQuAD Benchmark Dataset: Challenges and Lessons Learned
    Duer, Alexander
    Rauber, Andreas
    Filzmoser, Peter
    ADVANCES IN INFORMATION RETRIEVAL (ECIR 2018), 2018, 10772 : 102 - 113
  • [30] QookA: A Cooking Question Answering Dataset
    Frummet, Alexander
    Elsweiler, David
    PROCEEDINGS OF THE 2024 CONFERENCE ON HUMAN INFORMATION INTERACTION AND RETRIEVAL, CHIIR 2024, 2024, : 406 - 410