Cascaded deep neural network models for dialog state tracking

被引:0
|
作者
Yang, Guohua [1 ]
Wang, Xiaojie [2 ]
机构
[1] Beijing Univ Posts & Telecommun, Sch Comp Sci, Ctr Intelligence Sci & Technol, Beijing, Peoples R China
[2] Beijing Univ Posts & Telecommun, Sch Comp Sci, Beijing, Peoples R China
关键词
Dialog state tracking; Joint model; LSTM plus LSTM; CNN plus LSTM;
D O I
10.1007/s11042-018-6531-2
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Dialog state tracking (DST) maintains and updates dialog states at each time step as the dialog progresses. It is necessary to include dialog historical information in DST. Previous word-based DST models took historical utterances as a word sequence and used n-grams in the sequence as inputs of models. It suffered from the problem of data sparseness. This paper proposes a cascaded deep neural network framework for DST. It alleviates the problem of data sparseness by making use of the hierarchical structure in dialog. The bottom layer of the cascaded framework, implemented by an Long Short Term Memory (LSTM) or a Convolutional Neural Network (CNN), encodes the word sequence into a sentence embedding in each dialog turn, and the upper layer integrates the representation of each turn gradually to get the dialog state using an LSTM. The cascaded models integrate natural language understanding into DST, and the entire network is trained as a whole. The experimental results on the DSTC2 dataset indicate that the proposed models, LSTM+LSTM and CNN+LSTM, can achieve better performance than existing models.
引用
收藏
页码:9625 / 9643
页数:19
相关论文
共 50 条
  • [41] Knowledge-Based Dialog State Tracking
    Kadlec, Rudolf
    Vodolan, Miroslav
    Libovicky, Jindrich
    Macek, Jan
    Kleindienst, Jan
    2014 IEEE WORKSHOP ON SPOKEN LANGUAGE TECHNOLOGY SLT 2014, 2014, : 348 - 353
  • [42] Deep Neural Network Models for Colon Cancer Screening
    Kavitha, Muthu Subash
    Gangadaran, Prakash
    Jackson, Aurelia
    Maran, Balu Alagar Venmathi
    Kurita, Takio
    Ahn, Byeong-Cheol
    CANCERS, 2022, 14 (15)
  • [43] Individual differences among deep neural network models
    Mehrer, Johannes
    Spoerer, Courtney J.
    Kriegeskorte, Nikolaus
    Kietzmann, Tim C.
    NATURE COMMUNICATIONS, 2020, 11 (01)
  • [44] Predicted Robustness as QoS for Deep Neural Network Models
    Wang, Yue-Huan
    Li, Ze-Nan
    Xu, Jing-Wei
    Yu, Ping
    Chen, Taolue
    Ma, Xiao-Xing
    JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY, 2020, 35 (05) : 999 - 1015
  • [45] Individual differences among deep neural network models
    Johannes Mehrer
    Courtney J. Spoerer
    Nikolaus Kriegeskorte
    Tim C. Kietzmann
    Nature Communications, 11
  • [46] EvalDNN: A Toolbox for Evaluating Deep Neural Network Models
    Tian, Yongqiang
    Zeng, Zhihua
    Wen, Ming
    Liu, Yepang
    Kuo, Tzu-yang
    Cheung, Shing-Chi
    2020 ACM/IEEE 42ND INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING: COMPANION PROCEEDINGS (ICSE-COMPANION 2020), 2020, : 45 - 48
  • [47] Predicted Robustness as QoS for Deep Neural Network Models
    Yue-Huan Wang
    Ze-Nan Li
    Jing-Wei Xu
    Ping Yu
    Taolue Chen
    Xiao-Xing Ma
    Journal of Computer Science and Technology, 2020, 35 : 999 - 1015
  • [48] Window function convolution with deep neural network models
    Alkhanishvili, D.
    Porciani, C.
    Sefusatti, E.
    ASTRONOMY & ASTROPHYSICS, 2023, 669
  • [49] Deep problems with neural network models of human vision
    Bowers, Jeffrey S.
    Malhotra, Gaurav
    Dujmovic, Marin
    Llera Montero, Milton
    Tsvetkov, Christian
    Biscione, Valerio
    Puebla, Guillermo
    Adolfi, Federico
    Hummel, John E.
    Heaton, Rachel F.
    Evans, Benjamin D.
    Mitchell, Jeffrey
    Blything, Ryan
    BEHAVIORAL AND BRAIN SCIENCES, 2022, 46
  • [50] A note on factor normalization for deep neural network models
    Haobo Qi
    Jing Zhou
    Hansheng Wang
    Scientific Reports, 12