Cascaded deep neural network models for dialog state tracking

被引:0
|
作者
Yang, Guohua [1 ]
Wang, Xiaojie [2 ]
机构
[1] Beijing Univ Posts & Telecommun, Sch Comp Sci, Ctr Intelligence Sci & Technol, Beijing, Peoples R China
[2] Beijing Univ Posts & Telecommun, Sch Comp Sci, Beijing, Peoples R China
关键词
Dialog state tracking; Joint model; LSTM plus LSTM; CNN plus LSTM;
D O I
10.1007/s11042-018-6531-2
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Dialog state tracking (DST) maintains and updates dialog states at each time step as the dialog progresses. It is necessary to include dialog historical information in DST. Previous word-based DST models took historical utterances as a word sequence and used n-grams in the sequence as inputs of models. It suffered from the problem of data sparseness. This paper proposes a cascaded deep neural network framework for DST. It alleviates the problem of data sparseness by making use of the hierarchical structure in dialog. The bottom layer of the cascaded framework, implemented by an Long Short Term Memory (LSTM) or a Convolutional Neural Network (CNN), encodes the word sequence into a sentence embedding in each dialog turn, and the upper layer integrates the representation of each turn gradually to get the dialog state using an LSTM. The cascaded models integrate natural language understanding into DST, and the entire network is trained as a whole. The experimental results on the DSTC2 dataset indicate that the proposed models, LSTM+LSTM and CNN+LSTM, can achieve better performance than existing models.
引用
收藏
页码:9625 / 9643
页数:19
相关论文
共 50 条
  • [21] Adaptation and Contextualization of Deep Neural Network Models
    Kollias, Dimitrios
    Yu, Miao
    Tagaris, Athanasios
    Leontidis, Georgios
    Kollias, Stefanos
    Stafylopatis, Andreas
    2017 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI), 2017, : 1204 - 1211
  • [22] Sharing Deep Neural Network Models with Interpretation
    Wu, Huijun
    Wang, Chen
    Yin, Jie
    Lu, Kai
    Zhu, Liming
    WEB CONFERENCE 2018: PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE (WWW2018), 2018, : 177 - 186
  • [23] Human point-of-regard tracking using state space and modular neural network models
    Mitchell, JL
    Kothari, R
    NEURAL NETWORKS FOR SIGNAL PROCESSING VI, 1996, : 482 - 491
  • [24] Dialog State Tracking and Action Selection Using Deep Learning Mechanism for Interview Coaching
    Su, Ming-Hsiang
    Huang, Kun-Yi
    Yang, Tsung-Hsien
    Lai, Kuan-Jung
    Wu, Chung-Hsien
    PROCEEDINGS OF THE 2016 INTERNATIONAL CONFERENCE ON ASIAN LANGUAGE PROCESSING (IALP), 2016, : 6 - 9
  • [25] Dialog State Tracking Using Long Short-term Memory Neural Networks
    Yang, Xiaohao
    Liu, Jia
    16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5, 2015, : 1800 - 1804
  • [26] ROBUST DIALOG STATE TRACKING USING DELEXICALISED RECURRENT NEURAL NETWORKS AND UNSUPERVISED ADAPTATION
    Henderson, Matthew
    Thomson, Blaise
    Young, Steve
    2014 IEEE WORKSHOP ON SPOKEN LANGUAGE TECHNOLOGY SLT 2014, 2014, : 360 - 365
  • [27] An End-to-End Neural Dialog State Tracking for Task-Oriented Dialogs
    Kim, A-Yeong
    Kim, Tae-Hyeong
    Song, Hyun-Je
    Park, Seong-Bae
    2018 IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS (FUZZ-IEEE), 2018,
  • [28] Variational Hierarchical Dialog Autoencoder for Dialog State Tracking Data Augmentation
    Yoo, Kang Min
    Lee, Hanbit
    Dernoncourt, Franck
    Bui, Trung
    Chang, Walter
    Lee, Sang-Goo
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 3406 - 3425
  • [29] A multi-label cascaded neural network classification algorithm for automatic training and evolution of deep cascaded architecture
    Pakrashi, Arjun
    Mac Namee, Brian
    EXPERT SYSTEMS, 2021, 38 (07)
  • [30] Neural network surrogate models for equations of state
    Mentzer, Katherine L.
    Peterson, J. Luc
    PHYSICS OF PLASMAS, 2023, 30 (03)