RobustQA: Benchmarking the Robustness of Domain Adaptation for Open-Domain Question Answering

被引:0
|
作者
Han, Rujun [1 ]
Qi, Peng [1 ]
Zhang, Yuhao [1 ]
Liu, Lan [1 ]
Burger, Juliette [1 ]
Wang, William Yang [1 ]
Huang, Zhiheng [1 ]
Xiang, Bing [1 ]
Roth, Dan [1 ]
机构
[1] AWS AI Labs, Ann Arbor, MI 48108 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Open-domain question answering (ODQA) is a crucial task in natural language processing. A typical ODQA system relies on a retriever module to select relevant contexts from a large corpus for a downstream reading comprehension model. Existing ODQA datasets consist mainly of Wikipedia corpus, and are insufficient to study models' generalizability across diverse domains, as models are trained and evaluated on the same genre of data. We propose RobustQA(1), a novel benchmark consisting of datasets from 8 different domains, which facilitates the evaluation of ODQA's domain robustness. To build RobustQA, we annotate QA pairs in retrieval datasets with rigorous quality control. We further examine improving QA performances by incorporating unsupervised learning methods with target-domain corpus and adopting large generative language models. These methods can effectively improve model performances on RobustQA. However, experimental results demonstrate a significant gap from in-domain training, suggesting that RobustQA is a challenging benchmark to evaluate ODQA domain robustness.
引用
收藏
页码:4294 / 4311
页数:18
相关论文
共 50 条
  • [1] To Adapt or to Annotate: Challenges and Interventions for Domain Adaptation in Open-Domain Question Answering
    Dua, Dheeru
    Strubell, Emma
    Singh, Sameer
    Verga, Pat
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 14429 - 14446
  • [2] Advances in open-domain question answering
    Zhang, Zhi-Chang
    Zhang, Yu
    Liu, Ting
    Li, Sheng
    Tien Tzu Hsueh Pao/Acta Electronica Sinica, 2009, 37 (05): : 1058 - 1069
  • [3] Advances in question classification for open-domain question answering
    School of Computer Science and Technology, Anhui University of Technology, Maanshan
    Anhui
    243002, China
    不详
    Jiangsu
    210023, China
    Tien Tzu Hsueh Pao, 8 (1627-1636):
  • [4] Type checking in open-domain question answering
    Schlobach, S
    Olsthoorn, M
    de Rijke, M
    ECAI 2004: 16TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, PROCEEDINGS, 2004, 110 : 398 - 402
  • [5] Detrimental Contexts in Open-Domain Question Answering
    Oh, Philhoon
    Thorne, James
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 11589 - 11605
  • [6] Ranking and Sampling in Open-Domain Question Answering
    Xu, Yanfu
    Lin, Zheng
    Liu, Yuanxin
    Liu, Rui
    Wang, Weiping
    Meng, Dan
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 2412 - 2421
  • [7] Passage filtering for open-domain Question Answering
    Noguera, Elisa
    Llopis, Fernando
    Ferrandez, Antonio
    ADVANCES IN NATURAL LANGUAGE PROCESSING, PROCEEDINGS, 2006, 4139 : 534 - 540
  • [8] Open-domain textual question answering techniques
    Harabagiu, Sanda M.
    Maiorano, Steven J.
    Paşca, Marius A.
    Natural Language Engineering, 2003, 9 (03) : 231 - 267
  • [9] A Light Ranker for Open-Domain Question Answering
    Qiu, Boyu
    Xu, Jungang
    Chen, Xu
    Sun, Yingfei
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [10] PyGaggle: A Gaggle of Resources for Open-Domain Question Answering
    Pradeep, Ronak
    Chen, Haonan
    Gu, Lingwei
    Tamber, Manveer Singh
    Lin, Jimmy
    ADVANCES IN INFORMATION RETRIEVAL, ECIR 2023, PT III, 2023, 13982 : 148 - 162