Selecting Better Samples from Pre-trained LLMs: A Case Study on Question Generation

被引:0
|
作者
Yuan, Xingdi [1 ]
Wang, Tong [1 ]
Wang, Yen-Hsiang [2 ]
Fine, Emery [1 ]
Abdelgham, Rania [3 ]
Sauzeon, Helene [3 ]
Oudeyer, Pierre-Yves [3 ]
机构
[1] Microsoft Res, Montreal, PQ, Canada
[2] Natl Chung Hsing Univ, Taichung, Taiwan
[3] INRIA, Le Chesnay Rocquencourt, France
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Large Language Models (LLMs) have in recent years demonstrated impressive prowess in natural language generation. A common practice to improve generation diversity is to sample multiple outputs from the model. However, partly due to the inaccessibility of LLMs, there lacks a simple and robust way of selecting the best output from these stochastic samples. As a case study framed in the context of question generation, we propose two prompt-based approaches, namely round-trip and prompt-based score, to selecting high-quality questions from a set of LLM-generated candidates. Our method works without the need to modify the underlying model, nor does it rely on human-annotated references - both of which are realistic constraints for real-world deployment of LLMs. With automatic as well as human evaluations, we empirically demonstrate that our approach can effectively select questions of higher qualities than greedy generation. 1
引用
收藏
页码:12952 / 12965
页数:14
相关论文
共 50 条
  • [21] Leveraging pre-trained language models for code generation
    Soliman, Ahmed
    Shaheen, Samir
    Hadhoud, Mayada
    COMPLEX & INTELLIGENT SYSTEMS, 2024, 10 (03) : 3955 - 3980
  • [22] Synthesize, Prompt and Transfer: Zero-shot Conversational Question Generation with Pre-trained Language Model
    Zeng, Hongwei
    Wei, Bifan
    Liu, Jun
    Fu, Weiping
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 8989 - 9010
  • [23] PtbStolen: Pre-trained Encoder Stealing Through Perturbed Samples
    Zhang, Chuan
    Liang, Haotian
    Li, Zhuopeng
    Wu, Tong
    Wang, Licheng
    Zhu, Liehuang
    EMERGING INFORMATION SECURITY AND APPLICATIONS, EISA 2023, 2024, 2004 : 1 - 19
  • [24] From Cloze to Comprehension: Retrofitting Pre-trained Masked Language Models to Pre-trained Machine Reader
    Xu, Weiwen
    Li, Xin
    Zhang, Wenxuan
    Zhou, Meng
    Lam, Wai
    Si, Luo
    Bing, Lidong
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [25] From Dense to Sparse: Contrastive Pruning for Better Pre-trained Language Model Compression
    Xu, Runxin
    Luo, Fuli
    Wang, Chengyu
    Chang, Baobao
    Huang, Jun
    Huang, Songfang
    Huang, Fei
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 11547 - 11555
  • [26] Controllable Generation from Pre-trained Language Models via Inverse Prompting
    Zou, Xu
    Yin, Da
    Zhong, Qingyang
    Yang, Hongxia
    Yang, Zhilin
    Tang, Jie
    KDD '21: PROCEEDINGS OF THE 27TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2021, : 2450 - 2460
  • [27] Attribute Alignment: Controlling Text Generation from Pre-trained Language Models
    Yu, Dian
    Yu, Zhou
    Sagae, Kenji
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 2251 - 2268
  • [28] PluGeN: Multi-Label Conditional Generation from Pre-trained Models
    Wolczyk, Maciej
    Proszewska, Magdalena
    Maziarka, Lukasz
    Zieba, Maciej
    Wielopolski, Patryk
    Kurczab, Rafal
    Smieja, Marek
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 8647 - 8656
  • [29] Question-answering Forestry Pre-trained Language Model: ForestBERT
    Tan, Jingwei
    Zhang, Huaiqing
    Liu, Yang
    Yang, Jie
    Zheng, Dongping
    Linye Kexue/Scientia Silvae Sinicae, 2024, 60 (09): : 99 - 110
  • [30] Knowledge Grounded Pre-Trained Model For Dialogue Response Generation
    Wang, Yanmeng
    Rong, Wenge
    Zhang, Jianfei
    Ouyang, Yuanxin
    Xiong, Zhang
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,