Selecting Better Samples from Pre-trained LLMs: A Case Study on Question Generation

被引:0
|
作者
Yuan, Xingdi [1 ]
Wang, Tong [1 ]
Wang, Yen-Hsiang [2 ]
Fine, Emery [1 ]
Abdelgham, Rania [3 ]
Sauzeon, Helene [3 ]
Oudeyer, Pierre-Yves [3 ]
机构
[1] Microsoft Res, Montreal, PQ, Canada
[2] Natl Chung Hsing Univ, Taichung, Taiwan
[3] INRIA, Le Chesnay Rocquencourt, France
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Large Language Models (LLMs) have in recent years demonstrated impressive prowess in natural language generation. A common practice to improve generation diversity is to sample multiple outputs from the model. However, partly due to the inaccessibility of LLMs, there lacks a simple and robust way of selecting the best output from these stochastic samples. As a case study framed in the context of question generation, we propose two prompt-based approaches, namely round-trip and prompt-based score, to selecting high-quality questions from a set of LLM-generated candidates. Our method works without the need to modify the underlying model, nor does it rely on human-annotated references - both of which are realistic constraints for real-world deployment of LLMs. With automatic as well as human evaluations, we empirically demonstrate that our approach can effectively select questions of higher qualities than greedy generation. 1
引用
收藏
页码:12952 / 12965
页数:14
相关论文
共 50 条
  • [41] Sentiment Caption Generation from Visual Scene Using Pre-trained Language Model
    Zhang, Xiaochen
    Li, Jin
    Xu, Mengfan
    Li, Liangfu
    Guo, Longjiang
    Song, Yunpeng
    INTELLIGENT ROBOTICS AND APPLICATIONS, ICIRA 2024, PT VI, 2025, 15206 : 187 - 201
  • [42] Evaluating and Enhancing the Robustness of Code Pre-trained Models through Structure-Aware Adversarial Samples Generation
    Chen, Nuo
    Sun, Qiushi
    Wang, Jianing
    Gao, Ming
    Li, Xiaoli
    Li, Xiang
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 14857 - 14873
  • [43] A Comparative Study on Pre-Trained Models Based on BERT
    Zhang, Minghua
    2024 6TH INTERNATIONAL CONFERENCE ON NATURAL LANGUAGE PROCESSING, ICNLP 2024, 2024, : 326 - 330
  • [44] Pre-Trained Language-Meaning Models for Multilingual Parsing and Generation
    Wang, Chunliu
    Lai, Huiyuan
    Nissim, Malvina
    Bos, Johan
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 5586 - 5600
  • [45] Making Pre-trained Language Models Better Few-shot Learners
    Gao, Tianyu
    Fisch, Adam
    Chen, Danqi
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, 2021, : 3816 - 3830
  • [46] Better Few-Shot Text Classification with Pre-trained Language Model
    Chen, Zheng
    Zhang, Yunchen
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2021, PT II, 2021, 12892 : 537 - 548
  • [47] Commonsense Knowledge Reasoning and Generation with Pre-trained Language Models: A Survey
    Bhargava, Prajjwal
    Ng, Vincent
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 12317 - 12325
  • [48] AnchiBERT: A Pre-Trained Model for Ancient Chinese Language Understanding and Generation
    Tian, Huishuang
    Yang, Kexin
    Liu, Dayiheng
    Lv, Jiancheng
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [49] Leveraging Pre-Trained Language Model for Summary Generation on Short Text
    Zhao, Shuai
    You, Fucheng
    Liu, Zeng Yuan
    IEEE ACCESS, 2020, 8 : 228798 - 228803
  • [50] Non-Autoregressive Text Generation with Pre-trained Language Models
    Su, Yixuan
    Cai, Deng
    Wang, Yan
    Vandyke, David
    Baker, Simon
    Li, Piji
    Collier, Nigel
    16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 234 - 243