Learning from Mistakes: Combining Ontologies via Self-Training for Dialogue Generation

被引:0
|
作者
Reed, Lena [1 ]
Harrison, Vrindavan [1 ]
Oraby, Shereen [2 ]
Hakkani-Tur, Dilek [2 ]
Walker, Marilyn [1 ]
机构
[1] Univ Calif Santa Cruz, Nat Language & Dialogue Syst Lab, Santa Cruz, CA 95064 USA
[2] Amazon Alexa AI, Santa Cruz, CA USA
来源
SIGDIAL 2020: 21ST ANNUAL MEETING OF THE SPECIAL INTEREST GROUP ON DISCOURSE AND DIALOGUE (SIGDIAL 2020) | 2020年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Natural language generators (NLGs) for task-oriented dialogue typically take a meaning representation (MR) as input, and are trained end-to-end with a corpus of MR/utterance pairs, where the MRs cover a specific set of dialogue acts and domain attributes. Creation of such datasets is labor intensive and time consuming. Therefore, dialogue systems for new domain ontologies would benefit from using data for pre-existing ontologies. Here we explore, for the first time, whether it is possible to train an NLG for a new larger ontology using existing training sets for the restaurant domain, where each set is based on a different ontology. We create a new, larger combined ontology, and then train an NLG to produce utterances covering it. For example, if one dataset has attributes for family friendly and rating information, and the other has attributes for decor and service, our aim is an NLG for the combined ontology that can produce utterances that realize values for family friendly, rating, decor and service. Initial experiments with a baseline neural sequence-to-sequence model show that this task is surprisingly challenging. We then develop a novel self-training method that identifies (errorful) model outputs, automatically constructs a corrected MR input to form a new (MR, utterance) training pair, and then repeatedly adds these new instances back into the training data. We then test the resulting model on a new test set. The result is a self-trained model whose performance is an absolute 75.4% improvement over the baseline model. We also report a human qualitative evaluation of the final model showing that it achieves high naturalness, semantic coherence and grammaticality.
引用
收藏
页码:21 / 34
页数:14
相关论文
共 50 条
  • [1] State Value Generation with Prompt Learning and Self-Training for Low-Resource Dialogue State Tracking
    Gu, Ming
    Yang, Yan
    Chen, Chengcai
    Yu, Zhou
    ASIAN CONFERENCE ON MACHINE LEARNING, VOL 222, 2023, 222
  • [2] Combining Self-Training and Self-Supervised Learning for Unsupervised Disfluency Detection
    Wang, Shaolei
    Wang, Zhongyuan
    Che, Wanxiang
    Liu, Ting
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 1813 - 1822
  • [3] SELF-TRAINING CLASSIFIER VIA LOCAL LEARNING REGULARIZATION
    Cheng, Yong
    Zhao, Ruilian
    PROCEEDINGS OF 2009 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS, VOLS 1-6, 2009, : 454 - 459
  • [4] Improving Graph Neural Networks by combining active learning with self-training
    Katsimpras, Georgios
    Paliouras, Georgios
    DATA MINING AND KNOWLEDGE DISCOVERY, 2024, 38 (01) : 110 - 127
  • [5] Improving Graph Neural Networks by combining active learning with self-training
    Georgios Katsimpras
    Georgios Paliouras
    Data Mining and Knowledge Discovery, 2024, 38 : 110 - 127
  • [6] Unsupervised Learning of Object Landmarks via Self-Training Correspondence
    Mallis, Dimitrios
    Sanchez, Enrique
    Bell, Matt
    Tzimiropoulos, Georgios
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [7] Learning to Classify Skin Lesions via Self-Training and Self-Paced Learning
    Asare, Sarpong Kwadwo
    You, Fei
    Nartey, Obed Tettey
    2020 IEEE INTERNATIONAL CONFERENCE ON BIOINFORMATICS AND BIOMEDICINE, 2020, : 963 - 967
  • [8] Unsupervised Controllable Generation with Self-Training
    Chrysos, Grigorios G.
    Kossaifi, Jean
    Yu, Zhiding
    Anandkumar, Anima
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [9] Semi-Supervised Meta-Learning via Self-Training
    Zhou, Meng
    Li, Yaoyi
    Lu, Hongtao
    Cai Nengbin
    Zhao Xuejun
    2020 THE 3RD INTERNATIONAL CONFERENCE ON INTELLIGENT AUTONOMOUS SYSTEMS (ICOIAS'2020), 2020, : 1 - 7
  • [10] Learning from a Friend: Improving Event Extraction via Self-Training with Feedback from Abstract Meaning Representation
    Xu, Zhiyang
    Lee, Jay-Yoon
    Huang, Lifu
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 10421 - 10437