Improving generalisation to new speakers in spoken dialogue state tracking

被引:2
|
作者
Casanueva, Inigo [1 ]
Hain, Thomas [1 ]
Green, Phil [1 ]
机构
[1] Univ Sheffield, Dept Comp Sci, Sheffield, S Yorkshire, England
基金
英国工程与自然科学研究理事会;
关键词
dialogue state tracking; dysarthric speakers;
D O I
10.21437/Interspeech.2016-404
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Users with disabilities can greatly benefit from personalised voice-enabled environmental-control interfaces, but for users with speech impairments (e.g. dysarthria) poor ASR performance poses a challenge to successful dialogue. Statistical dialogue management has shown resilience against high ASR error rates, hence making it useful to improve the performance of these interfaces. However, little research was devoted to dialogue management personalisation to specific users so far. Recently, data driven discriminative models have been shown to yield the best performance in dialogue state tracking (the inference of the user goal from the dialogue history). However, due to the unique characteristics of each speaker, training a system for a new user when user specific data is not available can be challenging due to the mismatch between training and working conditions. This work investigates two methods to improve the performance with new speakers of a LSTM-based personalised state tracker: The use of speaker specific acoustic and ASR related features; and dropout regularisation. It is shown that in an environmental control system for dysarthric speakers, the combination of both techniques yields improvements of 3.5% absolute in state tracking accuracy. Further analysis explores the effect of using different amounts of speaker specific data to train the tracking system.
引用
收藏
页码:2726 / 2730
页数:5
相关论文
共 50 条
  • [11] Cognitive state classification in a spoken tutorial dialogue system
    Zhang, Tong
    Hasegawa-Johnson, Mark
    Levinson, Stephen E.
    SPEECH COMMUNICATION, 2006, 48 (06) : 616 - 632
  • [12] Evolvable dialogue state tracking for statistical dialogue management
    Kai YU
    Lu CHEN
    Kai SUN
    Qizhe XIE
    Su ZHU
    Frontiers of Computer Science, 2016, 10 (02) : 201 - 215
  • [13] Evolvable dialogue state tracking for statistical dialogue management
    Yu, Kai
    Chen, Lu
    Sun, Kai
    Xie, Qizhe
    Zhu, Su
    FRONTIERS OF COMPUTER SCIENCE, 2016, 10 (02) : 201 - 215
  • [14] Evolvable dialogue state tracking for statistical dialogue management
    Kai Yu
    Lu Chen
    Kai Sun
    Qizhe Xie
    Su Zhu
    Frontiers of Computer Science, 2016, 10 : 201 - 215
  • [15] Improving Dialogue State Tracking with Interactive Acts Attention and Attention Divergence Loss Function
    Lan, Jingjing
    Liang, Rongjiao
    Liu, Hai
    He, Chaobo
    Hao, Tianyong
    NEURAL COMPUTING FOR ADVANCED APPLICATIONS, NCAA 2024, PT III, 2025, 2183 : 116 - 131
  • [16] DSTEA: Improving Dialogue State Tracking via Entity Adaptive pre-training
    Lee, Yukyung
    Kim, Takyoung
    Yoon, Hoonsang
    Kang, Pilsung
    Bang, Junseong
    Kim, Misuk
    KNOWLEDGE-BASED SYSTEMS, 2024, 290
  • [17] Improving map generalisation with new pruning heuristics
    Taillandier, Patrick
    Gaffuri, Julien
    INTERNATIONAL JOURNAL OF GEOGRAPHICAL INFORMATION SCIENCE, 2012, 26 (07) : 1309 - 1323
  • [18] Towards Universal Dialogue State Tracking
    Ren, Liliang
    Xie, Kaige
    Chen, Lu
    Yu, Kai
    2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 2780 - 2786
  • [19] Dual Learning for Dialogue State Tracking
    Chen, Zhi
    Chen, Lu
    Zhao, Yanbin
    Zhu, Su
    Yu, Kai
    MAN-MACHINE SPEECH COMMUNICATION, NCMMSC 2022, 2023, 1765 : 293 - 305
  • [20] SCALABLE NEURAL DIALOGUE STATE TRACKING
    Balaraman, Vevake
    Magnini, Bernardo
    2019 IEEE AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING WORKSHOP (ASRU 2019), 2019, : 830 - 837