Knowledge-Grounded Dialogue Generation with Pre-trained Language Models

被引:0
|
作者
Zhao, Xueliang [1 ,2 ]
Wu, Wei [3 ]
Xu, Can [4 ]
Tao, Chongyang [4 ]
Zhao, Dongyan [1 ,2 ]
Yan, Rui [1 ,2 ,5 ]
机构
[1] Peking Univ, Wangxuan Inst Comp Technol, Beijing, Peoples R China
[2] Peking Univ, Ctr Data Sci, AAIS, Beijing, Peoples R China
[3] Meituan, Beijing, Peoples R China
[4] Microsoft Corp, Beijing, Peoples R China
[5] Beijing Acad Artificial Intelligence BAAI, Beijing, Peoples R China
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study knowledge-grounded dialogue generation with pre-trained language models. To leverage the redundant external knowledge under capacity constraint, we propose equipping response generation defined by a pre-trained language model with a knowledge selection module, and an unsupervised approach to jointly optimizing knowledge selection and response generation with unlabeled dialogues. Empirical results on two benchmarks indicate that our model can significantly outperform state-of-the-art methods in both automatic evaluation and human judgment.
引用
收藏
页码:3377 / 3390
页数:14
相关论文
共 50 条
  • [21] Adaptive Posterior Knowledge Selection for Improving Knowledge-Grounded Dialogue Generation
    Wang, Weichao
    Gao, Wei
    Feng, Shi
    Chen, Ling
    Wang, Daling
    International Conference on Information and Knowledge Management, Proceedings, 2021, : 1989 - 1998
  • [22] Probing Simile Knowledge from Pre-trained Language Models
    Chen, Weijie
    Chang, Yongzhu
    Zhang, Rongsheng
    Pu, Jiashu
    Chen, Guandan
    Zhang, Le
    Xi, Yadong
    Chen, Yijiang
    Su, Chang
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 5875 - 5887
  • [23] Efficient Latent Variable Modeling for Knowledge-Grounded Dialogue Generation
    Han, Gunsoo
    Jo, Daejin
    Nam, Daniel Wontae
    Yoon, Eunseop
    Kwon, Taehwan
    Rho, Seungeun
    On, Kyoung-Woon
    Yoo, Chang D.
    Kim, Sungwoong
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 2683 - 2702
  • [24] ProSide: Knowledge Projector and Sideway for Pre-trained Language Models
    He, Chaofan
    Lu, Gewei
    Shen, Liping
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT II, NLPCC 2024, 2025, 15360 : 56 - 68
  • [25] Continual knowledge infusion into pre-trained biomedical language models
    Jha, Kishlay
    Zhang, Aidong
    BIOINFORMATICS, 2022, 38 (02) : 494 - 502
  • [26] Exploring Pre-trained Language Models for Event Extraction and Generation
    Yang, Sen
    Feng, Dawei
    Qiao, Linbo
    Kan, Zhigang
    Li, Dongsheng
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 5284 - 5294
  • [27] STYLEDGPT: Stylized Response Generation with Pre-trained Language Models
    Yang, Ze
    Wu, Wei
    Xu, Can
    Liang, Xinnian
    Bai, Jiaqi
    Wang, Liran
    Wang, Wei
    Li, Zhoujun
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 1548 - 1559
  • [28] Scalable Educational Question Generation with Pre-trained Language Models
    Bulathwela, Sahan
    Muse, Hamze
    Yilmaz, Emine
    ARTIFICIAL INTELLIGENCE IN EDUCATION, AIED 2023, 2023, 13916 : 327 - 339
  • [29] TopicKS: Topic-driven Knowledge Selection for Knowledge-grounded Dialogue Generation
    Wang, Shiquan
    Si, Yuke
    Wei, Xiao
    Wang, Longbiao
    Zhuang, Zhiqiang
    Zhang, Xiaowang
    Dang, Jianwu
    INTERSPEECH 2022, 2022, : 1121 - 1125
  • [30] Pre-Trained Language Models and Their Applications
    Wang, Haifeng
    Li, Jiwei
    Wu, Hua
    Hovy, Eduard
    Sun, Yu
    ENGINEERING, 2023, 25 : 51 - 65