MULTIMODAL ADDRESSEE DETECTION IN MULTIPARTY DIALOGUE SYSTEMS

被引:0
|
作者
Tsai, T. J. [1 ]
Stolcke, Andreas [2 ]
Slaney, Malcolm [2 ]
机构
[1] Univ Calif Berkeley, Berkeley, CA 94720 USA
[2] Microsoft Res, Mountain View, CA USA
来源
2015 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING (ICASSP) | 2015年
关键词
addressee detection; dialog system; multimodality; multiparty; human-human-computer;
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Addressee detection answers the question, "Are you talking to me?" When multiple users interact with a dialogue system, it is important to know when a user is speaking to the computer and when he or she is speaking to another person. We approach this problem from a multimodal perspective, using lexical, acoustic, visual, dialog state, and beam-forming information. Using data from a multiparty dialogue system, we demonstrate the benefit of using multiple modalities over using a single modality. We also assess the relative importance of the various modalities in predicting the addressee. In our experiments, we find that acoustic features are by far the most important, that ASR and system-state information are useful, and that visual and beamforming features provide little additional benefit. Our study suggests that acoustic, lexical, and system state information are an effective, economical combination of modalities to use in addressee detection.
引用
收藏
页码:2314 / 2318
页数:5
相关论文
共 50 条
  • [21] Using Complexity-Identical Human- and Machine-Directed Utterances to Investigate Addressee Detection for Spoken Dialogue Systems
    Akhtiamov, Oleg
    Siegert, Ingo
    Karpov, Alexey
    Minker, Wolfgang
    SENSORS, 2020, 20 (09)
  • [22] A Gaze-based Method for Relating Group Involvement to Individual Engagement in Multimodal Multiparty Dialogue
    Oertel, Catharine
    Salvi, Giampiero
    ICMI'13: PROCEEDINGS OF THE 2013 ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2013, : 99 - 106
  • [23] Natural Multimodal Dialogue Systems: A Configurable Dialogue and Presentation Strategies Component
    Horchani, Meriam
    Caron, Benjamin
    Nigay, Laurence
    Panaget, Franck
    ICMI'07: PROCEEDINGS OF THE NINTH INTERNATIONAL CONFERENCE ON MULTIMODAL INTERFACES, 2007, : 291 - +
  • [24] The Analysis of the Evaluation Methodologies of Multimodal Dialogue Systems
    Lopez-Cozar, R.
    Callejas, Z.
    Gea, M.
    PROCESAMIENTO DEL LENGUAJE NATURAL, 2005, (34):
  • [25] Modeling context for referring in multimodal dialogue systems
    Landragin, F
    MODELING AND USING CONTEXT, PROCEEDINGS, 2005, 3554 : 240 - 253
  • [26] Knowledge-aware Multimodal Dialogue Systems
    Liao, Lizi
    Ma, Yunshan
    He, Xiangnan
    Hong, Richang
    Chua, Tat-Seng
    PROCEEDINGS OF THE 2018 ACM MULTIMEDIA CONFERENCE (MM'18), 2018, : 801 - 809
  • [27] Graphical and linguistic dialogue for intelligent multimodal systems
    Pineda, LA
    EXPERT SYSTEMS WITH APPLICATIONS, 1998, 14 (1-2) : 149 - 157
  • [28] Information seeking spoken dialogue systems - Part II: Multimodal dialogue
    Potamianos, Alexandros
    Fosler-Lussier, Eric
    Ammicht, Egbert
    Perakakis, Manolis
    IEEE TRANSACTIONS ON MULTIMEDIA, 2007, 9 (03) : 550 - 566
  • [29] Automatic evaluation tool for multimodal dialogue systems
    Wesseling, Hugo
    Bezold, Matthias
    Beringer, Nicole
    PERCEPTION IN MULTIMODAL DIALOGUE SYSTEMS, PROCEEDINGS, 2008, 5078 : 297 - 305
  • [30] Multimodal dialogue systems for interactive TV applications
    Ibrahim, A
    Johansson, P
    FOURTH IEEE INTERNATIONAL CONFERENCE ON MULTIMODAL INTERFACES, PROCEEDINGS, 2002, : 117 - 122