A Novel Multimodal Situated Spoken Dialog System for Human Robot Communication in Emergency Evacuation

被引:1
|
作者
Paul, Sheuli [1 ]
Sintek, Michael [2 ]
Silaghi, Marius [3 ]
Kepuska, Veton [3 ]
Robertson, Liam [4 ]
机构
[1] DRDC, Ottawa, ON, Canada
[2] DFKI GmbH, Kaiserslautern, Germany
[3] Florida Inst Technol, Melbourne, FL 32901 USA
[4] DND, Ottawa, ON, Canada
关键词
Axios" word used as WUW; MSSDS; MA; HRI; MDP;
D O I
10.1109/ICMLA55696.2022.00255
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Given the need for multimodal autonomous human-robot interactive systems serving complex situations, a noise-robust and contextaware situated multimodal spoken dialogue system (MSSDS) in emergency evacuation missions is presented. The MSSDS system is composed of: (1) a context and noise robust Wake-Up-Word (WUW) (a) to initiate the dialogue, and (b) to detect the context switching by explicit use of WUW during emergency evacuation mission steps, (2) multi-turn userinitiated interactive text-and-spoken dialog communication system, and (3) an interactive voice and text interface for human-robot communication designed based on dialogues used in real-world emergency situations. In an emergency environment, speech is mixed with different noises, and therefore communication using speech in such an environment is challenging. We handle the noise by using Team Connect Ceiling (TCC) beam-forming microphone arrays. Innovative and useful applications of spoken dialogue systems, presented as proof of concept, constitute another contribution. Numerous digital assistants, keyword spotting, and wake-up-words-based technologies have already been developed, but these are mainly used indoors. Our objective is to support communication in complex environments, e.g., indoors and outdoors, in human and machine teaming, via a wake-up-word-based multimodal interactive system. The development of the real-world application to communicate with the robot using multimodalities in complex situations based on the presented approach is in progress while the presented simulated approach is reflecting parts of this development. Numerous machine learning technologies and toolkits have been applied in this ongoing development process. The novelty of WUW-based MSSDS is discussed in this paper. Our Markov Decision Process (MDP) evaluation shows that the WUW-based MSSDS performs better.
引用
收藏
页码:1660 / 1665
页数:6
相关论文
共 50 条
  • [1] Situated language understanding for a spoken dialog system within vehicles
    Misu, Teruhisa
    Raux, Antoine
    Gupta, Rakesh
    Lane, Ian
    COMPUTER SPEECH AND LANGUAGE, 2015, 34 (01): : 186 - 200
  • [2] Leveraging Commonsense Reasoning and Multimodal Perception for Robot Spoken Dialog Systems
    Lu, Dongcai
    Zhang, Shiqi
    Stone, Peter
    Chen, Xiaoping
    2017 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2017, : 6582 - 6588
  • [3] Multimodal Spoken Dialog System Using State Estimation by Body Motion
    Koseki, Takeru
    Kosaka, Tetsuo
    2017 IEEE 6TH GLOBAL CONFERENCE ON CONSUMER ELECTRONICS (GCCE), 2017,
  • [4] A Multimodal Human-Robot-Dialog Applying Emotional Feedbacks
    Bannat, Alexander
    Blume, Juergen
    Geiger, Juergen T.
    Rehrl, Tobias
    Wallhoff, Frank
    Mayer, Christoph
    Radig, Bernd
    Sosnowski, Stefan
    Kuehnlenz, Kolja
    SOCIAL ROBOTICS, ICSR 2010, 2010, 6414 : 1 - 10
  • [5] Situated Communication for Joint Activity in Human-Robot Teams
    Kruijff, Geert-Jan M.
    Janicek, Miroslav
    Zender, Hendrik
    IEEE INTELLIGENT SYSTEMS, 2012, 27 (02) : 27 - 35
  • [6] Multimodal Resources for Human -Robot Communication Modelling
    Fotinea, Stavroula-Evita
    Efthimiou, Eleni
    Koutsombogera, Maria
    Dimou, Athanasia-Lida
    Goulas, Theodore
    Vasilaki, Kyriaki
    LREC 2016 - TENTH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2016, : 3455 - 3460
  • [7] Multimodal Strategies for Robot-to-Human Communication
    Donini, Massimo
    Gena, Cristina
    Mazzei, Alessandro
    COMPANION OF THE 2024 ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION, HRI 2024 COMPANION, 2024, : 417 - 421
  • [8] Situated Meaning in Multimodal Dialogue: Human-Robot and Human-Computer Interactions
    Pustejovsky, James
    Krishnaswamy, Nikhil
    TRAITEMENT AUTOMATIQUE DES LANGUES, 2020, 61 (03): : 17 - 41
  • [9] A Natural Conversational Virtual Human with Multimodal Dialog System
    Ali, Itimad Raheem
    Sulong, Ghazali
    Basori, Ahmad Hoirul
    JURNAL TEKNOLOGI, 2014, 71 (05):
  • [10] Novel Bidirectional Multimodal System for Affective Human-Robot Engagement
    Applewhite, Timothy
    Zhong, Vivienne Jia
    Dornberger, Rolf
    2021 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI 2021), 2021,