Multiparty Interaction Understanding Using Smart Multimodal Digital Signage

被引:13
|
作者
Tung, Tony [1 ,2 ]
Gomez, Randy [1 ,3 ]
Kawahara, Tatsuya
Matsuyama, Takashi [2 ]
机构
[1] Kyoto Univ, Acad Ctr Media Studies, Kyoto 6068501, Japan
[2] Kyoto Univ, Grad Sch Informat, Kyoto 6068501, Japan
[3] Honda Res Inst Japan, Wako, Saitama 6068501, Japan
关键词
Human-machine system; multimodal interaction dynamics; multiparty interaction; smart digital signage; RECOGNITION;
D O I
10.1109/THMS.2014.2326873
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper presents a novel multimodal system designed for multi-party human-human interaction analysis. The design of human-machine interfaces for multiple users is challenging because simultaneous processing of actions and reactions have to be consistent. The proposed system consists of a large display equipped with multiple sensing devices: microphone array, HD video cameras, and depth sensors. Multiple users positioned in front of the panel freely interact using voice or gesture while looking at the displayed content, without wearing any particular devices (such as motion capture sensors or head mounted devices). Acoustic and visual information is captured and processed jointly using established and state-of-the-art techniques to obtain individual speech and gaze direction. Furthermore, a new framework is proposed to model A/V multimodal interaction between verbal and nonverbal communication events. Dynamics of audio signals obtained from speaker diarization and head poses extracted from video images are modeled using hybrid dynamical systems (HDS). We show that HDS temporal structure characteristics can be used for multimodal interaction level estimation, which is useful feedback that can help to improve multi-party communication experience. Experimental results using synthetic and real-world datasets of group communication such as poster presentations show the feasibility of the proposed multimodal system.
引用
收藏
页码:625 / 637
页数:13
相关论文
共 50 条
  • [41] Smart Home: Multimodal Interaction for Control of Home Devices
    Angel Contreras-Castaneda, Miguel
    Antonio Holgado-Terriza, Juan
    Pomboza-Junez, Gonzalo
    Paderewski-Rodriguez, Patricia
    Luis Gutierrez-Vela, Francisco
    PROCEEDINGS OF THE XX INTERNATIONAL CONFERENCE ON HUMAN-COMPUTER INTERACTION (INTERACCION'2019), 2019,
  • [42] Exploring Smart Agents for the Interaction with Multimodal Mediated Environments
    Richer, Robert
    Zhao, Nan
    Eskofier, Bjoern M.
    Paradiso, Joseph A.
    MULTIMODAL TECHNOLOGIES AND INTERACTION, 2020, 4 (02) : 1 - 18
  • [43] An innovative framework to support multimodal interaction with Smart Environments
    Gabbanini, Francesco
    Burzagli, Laura
    Emiliani, Pier Luigi
    EXPERT SYSTEMS WITH APPLICATIONS, 2012, 39 (03) : 2239 - 2246
  • [44] An Accessible Smart Home Based on Integrated Multimodal Interaction
    Rocha, Ana Patricia
    Ketsmur, Maksym
    Almeida, Nuno
    Teixeira, Antonio
    SENSORS, 2021, 21 (16)
  • [45] A Quantum -Like multimodal network framework for modeling interaction dynamics in multiparty conversational sentiment analysis
    Zhang, Yazhou
    Song, Dawei
    Li, Xiang
    Zhang, Peng
    Wang, Panpan
    Rong, Lu
    Yu, Guangliang
    Wang, Bo
    INFORMATION FUSION, 2020, 62 : 14 - 31
  • [46] Digital Signage Personalized Recommendation Service Using loT Based DID
    Kim, Wonseok
    Bak, Eunsik
    Choi, Euiin
    2017 6TH INTERNATIONAL CONFERENCE ON APPLIED SOCIAL SCIENCE (ICASS 2017), PT 1, 2017, 97 : 256 - 260
  • [47] Presentation Technique of Scents Using Mobile Olfactory Display for Digital Signage
    Sugimoto, Sayumi
    Segawa, Ryo
    Noguchi, Daisuke
    Bannai, Yuichi
    Okada, Kenichi
    HUMAN-COMPUTER INTERACTION - INTERACT 2011, PT III, 2011, 6948 : 323 - 337
  • [48] A Study on Smart IoT Hub for Intelligent Signage Services Using Trust Information
    Lee, Jun Wook
    Kim, Yong Woo
    2018 32ND INTERNATIONAL CONFERENCE ON INFORMATION NETWORKING (ICOIN), 2018, : 76 - 79
  • [49] Human-Robot Collaborative Tutoring using Multiparty Multimodal Spoken Dialogue
    Al Moubayed, Samer
    Beskow, Jonas
    Bollepalli, Bajibabu
    Gustafson, Joakim
    Hussen-Abdelaziz, Ahmed
    Johansson, Martin
    Koutsombogera, Maria
    Lopes, Jose David
    Novikova, Jekaterina
    Oertel, Catharine
    Skantze, Gabriel
    Stefanov, Kalin
    Varol, Gul
    HRI'14: PROCEEDINGS OF THE 2014 ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION, 2014, : 112 - 113
  • [50] Multimodal digital semiotics: the interaction of language with other resources
    O'Halloran, Kay L.
    Marissa, K. L. E.
    Podlasov, Alexey
    Tan, Sabine
    TEXT & TALK, 2013, 33 (4-5) : 665 - 690