Selecting Better Samples from Pre-trained LLMs: A Case Study on Question Generation

被引:0
|
作者
Yuan, Xingdi [1 ]
Wang, Tong [1 ]
Wang, Yen-Hsiang [2 ]
Fine, Emery [1 ]
Abdelgham, Rania [3 ]
Sauzeon, Helene [3 ]
Oudeyer, Pierre-Yves [3 ]
机构
[1] Microsoft Res, Montreal, PQ, Canada
[2] Natl Chung Hsing Univ, Taichung, Taiwan
[3] INRIA, Le Chesnay Rocquencourt, France
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Large Language Models (LLMs) have in recent years demonstrated impressive prowess in natural language generation. A common practice to improve generation diversity is to sample multiple outputs from the model. However, partly due to the inaccessibility of LLMs, there lacks a simple and robust way of selecting the best output from these stochastic samples. As a case study framed in the context of question generation, we propose two prompt-based approaches, namely round-trip and prompt-based score, to selecting high-quality questions from a set of LLM-generated candidates. Our method works without the need to modify the underlying model, nor does it rely on human-annotated references - both of which are realistic constraints for real-world deployment of LLMs. With automatic as well as human evaluations, we empirically demonstrate that our approach can effectively select questions of higher qualities than greedy generation. 1
引用
收藏
页码:12952 / 12965
页数:14
相关论文
共 50 条
  • [31] Data Redaction from Pre-trained GANs
    Kong, Zhifeng
    Chaudhuri, Kamalika
    2023 IEEE CONFERENCE ON SECURE AND TRUSTWORTHY MACHINE LEARNING, SATML, 2023, : 638 - 677
  • [32] IndicBART: A Pre-trained Model for Indic Natural Language Generation
    Dabre, Raj
    Shrotriya, Himani
    Kunchukuttan, Anoop
    Puduppully, Ratish
    Khapra, Mitesh M.
    Kumar, Pratyush
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), 2022, : 1849 - 1863
  • [33] Exploring Pre-trained Language Models for Event Extraction and Generation
    Yang, Sen
    Feng, Dawei
    Qiao, Linbo
    Kan, Zhigang
    Li, Dongsheng
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 5284 - 5294
  • [34] Automatic question-answer pairs generation using pre-trained large language models in higher education
    Ling J.
    Afzaal M.
    Computers and Education: Artificial Intelligence, 2024, 6
  • [35] STYLEDGPT: Stylized Response Generation with Pre-trained Language Models
    Yang, Ze
    Wu, Wei
    Xu, Can
    Liang, Xinnian
    Bai, Jiaqi
    Wang, Liran
    Wang, Wei
    Li, Zhoujun
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 1548 - 1559
  • [36] StyleCineGAN: Landscape Cinemagraph Generation using a Pre-trained StyleGAN
    Choi, Jongwoo
    Seo, Kwanggyoon
    Ashtari, Amirsaman
    Noh, Junyong
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2024, 2024, : 7872 - 7881
  • [37] On the Transferability of Pre-trained Language Models: A Study from Artificial Datasets
    Chiang, Cheng-Han
    Lee, Hung-yi
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 10518 - 10525
  • [38] Pre-trained Language Models for the Legal Domain: A Case Study on Indian Law
    Paul, Shounak
    Mandal, Arpan
    Goyal, Pawan
    Ghosh, Saptarshi
    PROCEEDINGS OF THE 19TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND LAW, ICAIL 2023, 2023, : 187 - 196
  • [39] A Statistical Language Model for Pre-Trained Sequence Labeling: A Case Study on Vietnamese
    Liao, Xianwen
    Huang, Yongzhong
    Yang, Peng
    Chen, Lei
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2022, 21 (03)
  • [40] A Pre-trained Language Model for Medical Question Answering Based on Domain Adaption
    Liu, Lang
    Ren, Junxiang
    Wu, Yuejiao
    Song, Ruilin
    Cheng, Zhen
    Wang, Sibo
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, NLPCC 2022, PT II, 2022, 13552 : 216 - 227