Cooperative Self-training of Machine Reading Comprehension

被引:0
|
作者
Luo, Hongyin [1 ]
Li, Shang-Wen [2 ]
Gao, Mingye [3 ]
Yu, Seunghak [2 ]
Glass, James [1 ]
机构
[1] MIT CSAIL, Cambridge, MA 02139 USA
[2] Amazon AI, Bellevue, WA USA
[3] MIT MTL, Cambridge, MA USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pretrained language models have significantly improved the performance of downstream language understanding tasks, including extractive question answering, by providing high-quality contextualized word embeddings. However, training question answering models still requires large amounts of annotated data for specific domains. In this work, we propose a cooperative self-training framework, RGX, for automatically generating more non-trivial question-answer pairs to improve model performance. RGX is built upon a masked answer extraction task with an interactive learning environment containing an answer entity Recognizer, a question Generator, and an answer eXtractor. Given a passage with a masked entity, the generator generates a question around the entity, and the extractor is trained to extract the masked entity with the generated question and raw texts. The framework allows the training of question generation and answering models on any text corpora without annotation. We further leverage a self-training technique to improve the performance of both question generation and answer extraction models. Experiment results show that RGX outperforms the state-of-the-art (SOTA) pretrained language models and transfer learning approaches on standard question-answering benchmarks, and yields the new SOTA performance under given model size and transfer learning settings.
引用
收藏
页码:244 / 257
页数:14
相关论文
共 50 条
  • [31] Adversarial self-training for robustness and generalization
    Li, Zhuorong
    Wu, Minghui
    Jin, Canghong
    Yu, Daiwei
    Yu, Hongchuan
    PATTERN RECOGNITION LETTERS, 2024, 185 : 117 - 123
  • [32] Unsupervised Controllable Generation with Self-Training
    Chrysos, Grigorios G.
    Kossaifi, Jean
    Yu, Zhiding
    Anandkumar, Anima
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [33] Self-training for Cell Segmentation and Counting
    Luo, J.
    Oore, S.
    Hollensen, P.
    Fine, A.
    Trappenberg, T.
    ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, 11489 : 406 - 412
  • [34] CONSIDERATIONS ON SELF-TRAINING IN THE INNOVATION UNION
    Blaga, Petruta
    Tripon, Avram
    STUDIES ON LITERATURE, DISCOURSE AND MULTICULTURAL DIALOGUE: COMMUNICATION AND PUBLIC RELATIONS, 2013, : 56 - 61
  • [35] Reranking and Self-Training for Parser Adaptation
    McClosky, David
    Charniak, Eugene
    Johnson, Mark
    COLING/ACL 2006, VOLS 1 AND 2, PROCEEDINGS OF THE CONFERENCE, 2006, : 337 - 344
  • [36] Crafting networks: A self-training intervention
    Wang, Huatian
    Demerouti, Evangelia
    Rispens, Sonja
    van Gool, Piet
    JOURNAL OF VOCATIONAL BEHAVIOR, 2024, 149
  • [37] Adaptive Self-Training for Object Detection
    Vandeghen, Renaud
    Louppe, Gilles
    Van Droogenbroeck, Marc
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 914 - 923
  • [38] Self-training of Residents in the Specialization Process
    Gonzalez Mesa, Maria Isabel
    Zerquera Alvarez, Carlos Esteban
    Machin Asia, Annia
    MEDISUR-REVISTA DE CIENCIAS MEDICAS DE CIENFUEGOS, 2014, 12 (01): : 329 - 333
  • [39] Discriminative Self-training for Punctuation Prediction
    Chen, Qian
    Wang, Wen
    Chen, Mengzhe
    Zhang, Qinglin
    INTERSPEECH 2021, 2021, : 771 - 775
  • [40] Self-Training for Unsupervised Parsing with PRPN
    Mohananey, Anhad
    Kann, Katharina
    Bowman, Samuel R.
    16TH INTERNATIONAL CONFERENCE ON PARSING TECHNOLOGIES AND IWPT 2020 SHARED TASK ON PARSING INTO ENHANCED UNIVERSAL DEPENDENCIES, 2020, : 105 - 110