Learning from Mistakes: Combining Ontologies via Self-Training for Dialogue Generation

被引:0
|
作者
Reed, Lena [1 ]
Harrison, Vrindavan [1 ]
Oraby, Shereen [2 ]
Hakkani-Tur, Dilek [2 ]
Walker, Marilyn [1 ]
机构
[1] Univ Calif Santa Cruz, Nat Language & Dialogue Syst Lab, Santa Cruz, CA 95064 USA
[2] Amazon Alexa AI, Santa Cruz, CA USA
来源
SIGDIAL 2020: 21ST ANNUAL MEETING OF THE SPECIAL INTEREST GROUP ON DISCOURSE AND DIALOGUE (SIGDIAL 2020) | 2020年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Natural language generators (NLGs) for task-oriented dialogue typically take a meaning representation (MR) as input, and are trained end-to-end with a corpus of MR/utterance pairs, where the MRs cover a specific set of dialogue acts and domain attributes. Creation of such datasets is labor intensive and time consuming. Therefore, dialogue systems for new domain ontologies would benefit from using data for pre-existing ontologies. Here we explore, for the first time, whether it is possible to train an NLG for a new larger ontology using existing training sets for the restaurant domain, where each set is based on a different ontology. We create a new, larger combined ontology, and then train an NLG to produce utterances covering it. For example, if one dataset has attributes for family friendly and rating information, and the other has attributes for decor and service, our aim is an NLG for the combined ontology that can produce utterances that realize values for family friendly, rating, decor and service. Initial experiments with a baseline neural sequence-to-sequence model show that this task is surprisingly challenging. We then develop a novel self-training method that identifies (errorful) model outputs, automatically constructs a corrected MR input to form a new (MR, utterance) training pair, and then repeatedly adds these new instances back into the training data. We then test the resulting model on a new test set. The result is a self-trained model whose performance is an absolute 75.4% improvement over the baseline model. We also report a human qualitative evaluation of the final model showing that it achieves high naturalness, semantic coherence and grammaticality.
引用
收藏
页码:21 / 34
页数:14
相关论文
共 50 条
  • [31] ADAPTING SOUND RECOGNITION TO A NEW ENVIRONMENT VIA SELF-TRAINING
    Ebbers, Janek
    Keyser, Moritz Curt
    Haeb-Umbach, Reinhold
    29TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO 2021), 2021, : 1135 - 1139
  • [32] Weakly-Supervised Semantic Segmentation via Self-training
    Cheng, Hao
    Gu, Chaochen
    Wu, Kaijie
    2020 4TH INTERNATIONAL CONFERENCE ON CONTROL ENGINEERING AND ARTIFICIAL INTELLIGENCE (CCEAI 2020), 2020, 1487
  • [33] Soil Temperature Prediction via Self-Training: Izmir Case
    Tuysuzoglu, Goksu
    Birant, Derya
    Kiranoglu, Volkan
    JOURNAL OF AGRICULTURAL SCIENCES-TARIM BILIMLERI DERGISI, 2022, 28 (01): : 47 - 62
  • [34] Combining Coregularization and Consensus-based Self-Training for Multilingual Text Categorization
    Amini, Massih-Reza
    Goutte, Cyril
    Usunier, Nicolas
    SIGIR 2010: PROCEEDINGS OF THE 33RD ANNUAL INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH DEVELOPMENT IN INFORMATION RETRIEVAL, 2010, : 475 - 482
  • [35] Multi-Task Self-Training for Learning General Representations
    Ghiasi, Golnaz
    Zoph, Barret
    Cubuk, Ekin D.
    Quoc V Le
    Lin, Tsung-Yi
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 8836 - 8845
  • [36] Training kindergarten children on learning from their mistakes
    Ger, Ebru
    Roebers, Claudia M.
    BRITISH JOURNAL OF DEVELOPMENTAL PSYCHOLOGY, 2024, 42 (04) : 478 - 494
  • [37] Neighborhood-Regularized Self-Training for Learning with Few Labels
    Xu, Ran
    Yu, Yue
    Cui, Hejie
    Kan, Xuan
    Zhu, Yanqiao
    Ho, Joyce
    Zhang, Chao
    Yang, Carl
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 9, 2023, : 10611 - 10619
  • [38] Deep unsupervised shadow detection with curriculum learning and self-training
    Zhang, Qiang
    Guo, Hongyuan
    Li, Guanghe
    Zhang, Tianlu
    Jiao, Qiang
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2024, 248
  • [39] Dynamic self-training with less uncertainty for graph imbalance learning
    Juan, Xin
    Peng, Meixin
    Wang, Xin
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 271
  • [40] Semi-supervised Continual Learning with Meta Self-training
    Ho, Stella
    Liu, Ming
    Du, Lan
    Li, Yunfeng
    Gao, Longxiang
    Gao, Shang
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 4024 - 4028