Learning from Mistakes: Combining Ontologies via Self-Training for Dialogue Generation

被引:0
|
作者
Reed, Lena [1 ]
Harrison, Vrindavan [1 ]
Oraby, Shereen [2 ]
Hakkani-Tur, Dilek [2 ]
Walker, Marilyn [1 ]
机构
[1] Univ Calif Santa Cruz, Nat Language & Dialogue Syst Lab, Santa Cruz, CA 95064 USA
[2] Amazon Alexa AI, Santa Cruz, CA USA
来源
SIGDIAL 2020: 21ST ANNUAL MEETING OF THE SPECIAL INTEREST GROUP ON DISCOURSE AND DIALOGUE (SIGDIAL 2020) | 2020年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Natural language generators (NLGs) for task-oriented dialogue typically take a meaning representation (MR) as input, and are trained end-to-end with a corpus of MR/utterance pairs, where the MRs cover a specific set of dialogue acts and domain attributes. Creation of such datasets is labor intensive and time consuming. Therefore, dialogue systems for new domain ontologies would benefit from using data for pre-existing ontologies. Here we explore, for the first time, whether it is possible to train an NLG for a new larger ontology using existing training sets for the restaurant domain, where each set is based on a different ontology. We create a new, larger combined ontology, and then train an NLG to produce utterances covering it. For example, if one dataset has attributes for family friendly and rating information, and the other has attributes for decor and service, our aim is an NLG for the combined ontology that can produce utterances that realize values for family friendly, rating, decor and service. Initial experiments with a baseline neural sequence-to-sequence model show that this task is surprisingly challenging. We then develop a novel self-training method that identifies (errorful) model outputs, automatically constructs a corrected MR input to form a new (MR, utterance) training pair, and then repeatedly adds these new instances back into the training data. We then test the resulting model on a new test set. The result is a self-trained model whose performance is an absolute 75.4% improvement over the baseline model. We also report a human qualitative evaluation of the final model showing that it achieves high naturalness, semantic coherence and grammaticality.
引用
收藏
页码:21 / 34
页数:14
相关论文
共 50 条
  • [21] Adversarial Domain Adaptation Enhanced via Self-training
    Altinel, Fazil
    Akkaya, Ibrahim Batuhan
    29TH IEEE CONFERENCE ON SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS (SIU 2021), 2021,
  • [22] SLADE: A Self-Training Framework For Distance Metric Learning
    Duan, Jiali
    Lin, Yen-Liang
    Son Tran
    Davis, Larry S.
    Kuo, C-C Jay
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 9639 - 9648
  • [23] Cost Sensitive Active Learning Based on Self-training
    Wu, Yongcheng
    PROCEEDINGS OF 2014 IEEE INTERNATIONAL CONFERENCE ON PROGRESS IN INFORMATICS AND COMPUTING (PIC), 2014, : 42 - 45
  • [24] Self-Training with Contrastive Learning for Adversarial Domain Adaptation
    Zhang, Xingyi (xyzhanghust@gmail.com), 1600, Institute of Electrical and Electronics Engineers Inc.
  • [25] Text Classification Based on Transfer Learning and Self-Training
    Zheng, Yabin
    Teng, Shaohua
    Liu, Zhiyuan
    Sun, Maosong
    ICNC 2008: FOURTH INTERNATIONAL CONFERENCE ON NATURAL COMPUTATION, VOL 3, PROCEEDINGS, 2008, : 363 - 367
  • [26] Combining instance selection and self-training to improve data stream quantification
    Maletzke A.G.
    dos Reis D.M.
    Batista G.E.A.P.A.
    Journal of the Brazilian Computer Society, 2018, 24 (01)
  • [27] Combining Curriculum Learning and Knowledge Distillation for Dialogue Generation
    Zhu, Qingqing
    Chen, Xiuying
    Wu, Pengfei
    Liu, JunFei
    Zhao, Dongyan
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 1284 - 1295
  • [28] Combining Semantic Self-Supervision and Self-Training for Domain Adaptation in Semantic Segmentation
    Niemeijer, Joshua
    Schaefer, Joerg P.
    2021 IEEE INTELLIGENT VEHICLES SYMPOSIUM WORKSHOPS (IV WORKSHOPS), 2021, : 364 - 371
  • [29] Automatic generation of spoken dialogue from medical plans and ontologies
    Beveridge, Martin
    Fox, John
    JOURNAL OF BIOMEDICAL INFORMATICS, 2006, 39 (05) : 482 - 499
  • [30] Distantly Supervised Biomedical Relation Extraction via Negative Learning and Noisy Student Self-Training
    Dai, Yuanfei
    Zhang, Bin
    Wang, Shiping
    IEEE-ACM TRANSACTIONS ON COMPUTATIONAL BIOLOGY AND BIOINFORMATICS, 2024, 21 (06) : 1697 - 1708