Selecting Better Samples from Pre-trained LLMs: A Case Study on Question Generation

被引:0
|
作者
Yuan, Xingdi [1 ]
Wang, Tong [1 ]
Wang, Yen-Hsiang [2 ]
Fine, Emery [1 ]
Abdelgham, Rania [3 ]
Sauzeon, Helene [3 ]
Oudeyer, Pierre-Yves [3 ]
机构
[1] Microsoft Res, Montreal, PQ, Canada
[2] Natl Chung Hsing Univ, Taichung, Taiwan
[3] INRIA, Le Chesnay Rocquencourt, France
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Large Language Models (LLMs) have in recent years demonstrated impressive prowess in natural language generation. A common practice to improve generation diversity is to sample multiple outputs from the model. However, partly due to the inaccessibility of LLMs, there lacks a simple and robust way of selecting the best output from these stochastic samples. As a case study framed in the context of question generation, we propose two prompt-based approaches, namely round-trip and prompt-based score, to selecting high-quality questions from a set of LLM-generated candidates. Our method works without the need to modify the underlying model, nor does it rely on human-annotated references - both of which are realistic constraints for real-world deployment of LLMs. With automatic as well as human evaluations, we empirically demonstrate that our approach can effectively select questions of higher qualities than greedy generation. 1
引用
收藏
页码:12952 / 12965
页数:14
相关论文
共 50 条
  • [1] Are Pre-trained Convolutions Better than Pre-trained Transformers?
    Tay, Yi
    Dehghani, Mostafa
    Gupta, Jai
    Aribandi, Vamsi
    Bahri, Dara
    Qin, Zhen
    Metzler, Donald
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, 2021, : 4349 - 4359
  • [2] Conditional pre-trained attention based Chinese question generation
    Zhang, Liang
    Fang, Ligang
    Fan, Zheng
    Li, Wei
    An, Jing
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2021, 33 (20):
  • [3] Scalable Educational Question Generation with Pre-trained Language Models
    Bulathwela, Sahan
    Muse, Hamze
    Yilmaz, Emine
    ARTIFICIAL INTELLIGENCE IN EDUCATION, AIED 2023, 2023, 13916 : 327 - 339
  • [4] Can LLMs Facilitate Interpretation of Pre-trained Language Models?
    Mousi, Basel
    Durrani, Nadir
    Dalvi, Fahim
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 3248 - 3268
  • [5] An Extensive Study on Pre-trained Models for Program Understanding and Generation
    Zeng, Zhengran
    Ta, Hanzhuo
    Zhang, Haotian
    Li, Jing
    Zhang, Yuqun
    Zhang, Lingming
    PROCEEDINGS OF THE 31ST ACM SIGSOFT INTERNATIONAL SYMPOSIUM ON SOFTWARE TESTING AND ANALYSIS, ISSTA 2022, 2022, : 39 - 51
  • [6] UniRaG: Unification, Retrieval, and Generation for Multimodal Question Answering With Pre-Trained Language Models
    Lim, Qi Zhi
    Lee, Chin Poo
    Lim, Kian Ming
    Samingan, Ahmad Kamsani
    IEEE ACCESS, 2024, 12 : 71505 - 71519
  • [7] Pre-trained Language Model for Biomedical Question Answering
    Yoon, Wonjin
    Lee, Jinhyuk
    Kim, Donghyeon
    Jeong, Minbyul
    Kang, Jaewoo
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2019, PT II, 2020, 1168 : 727 - 740
  • [8] Math-LLMs: AI Cyberinfrastructure with Pre-trained Transformers for Math Education
    Zhang, Fan
    Li, Chenglu
    Henkel, Owen
    Xing, Wanli
    Baral, Sami
    Heffernan, Neil
    Li, Hai
    INTERNATIONAL JOURNAL OF ARTIFICIAL INTELLIGENCE IN EDUCATION, 2024,
  • [9] Towards automatic question generation using pre-trained model in academic field for Bahasa Indonesia
    Suhartono, Derwin
    Majiid, Muhammad Rizki Nur
    Fredyan, Renaldy
    EDUCATION AND INFORMATION TECHNOLOGIES, 2024, 29 (16) : 21295 - 21330
  • [10] Multi-Label Conditional Generation From Pre-Trained Models
    Proszewska, Magdalena
    Wolczyk, Maciej
    Zieba, Maciej
    Wielopolski, Patryk
    Maziarka, Lukasz
    Smieja, Marek
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (09) : 6185 - 6198