Comparing Single-modal and Multimodal Interaction in an Augmented Reality System

被引:0
|
作者
Wang, Zhimin [1 ,2 ]
Yu, Huangyue [1 ]
Wang, Haofei [2 ]
Wang, Zongji [1 ]
Lu, Feng [1 ,2 ]
机构
[1] Beihang Univ, Sch CSE, State Key Lab VR Technol & Syst, Beijing, Peoples R China
[2] Peng Cheng Lab, Shenzhen, Peoples R China
基金
中国国家自然科学基金;
关键词
multimodal interaction; augmented reality; gaze; gesture; speech; AR system;
D O I
10.1109/ISMAR-Adjunct51615.2020.00052
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multimodal interaction is expected to offer better user experience in Augmented Reality (AR), and thus becomes a recent research focus. However, due to the lack of hardware-level support, most existing works only combine two modalities at a time, e.g., gesture and speech. Gaze-based interaction techniques have been explored for the screen-based application, but rarely been used in AR systemsy configurable augmented reality system. In this paper, we propose a multimodal interactive system that integrates gaze, gesture and speech in a flexibly configurable augmented reality system. Our lightweight head-mounted device supports accurate gaze tracking, hand gesture recognition and speech recognition simultaneously. More importantly, the system can be easily configured into different modality combinations to study the effects of different interaction techniques. We evaluated the system in the table lamps scenario, and compared the performance of different interaction techniques. The experimental results show that the Gaze+Gesture+Speech is superior in terms of performance.
引用
收藏
页码:165 / 166
页数:2
相关论文
共 50 条
  • [1] A Wheelchair Control System Using Human-Machine Interaction: Single-Modal and Multimodal Approaches
    Shahin, Mohamed K.
    Tharwat, Alaa
    Gaber, Tarek
    Hassanien, Aboul Ella
    JOURNAL OF INTELLIGENT SYSTEMS, 2019, 28 (01) : 115 - 132
  • [2] Multimodal interaction with a wearable augmented reality system
    Kölsch, M
    Bane, R
    Höllerer, T
    Turk, M
    IEEE COMPUTER GRAPHICS AND APPLICATIONS, 2006, 26 (03) : 62 - 71
  • [3] Multimodal Interaction in Augmented Reality
    Chen, Zhaorui
    Li, Jinzhou
    Hua, Yifan
    Shen, Rui
    Basu, Anup
    2017 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2017, : 206 - 209
  • [4] Dominant SIngle-Modal SUpplementary Fusion (SIMSUF) for Multimodal Sentiment Analysis
    Huang, Jian
    Ji, Yanli
    Qin, Zhen
    Yang, Yang
    Shen, Heng Tao
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 8383 - 8394
  • [5] Co-Training for Deep Object Detection: Comparing Single-Modal and Multi-Modal Approaches
    Gomez, Jose L.
    Villalonga, Gabriel
    Lopez, Antonio M.
    SENSORS, 2021, 21 (09)
  • [6] Colour-Touch Cross-Modal Correspondence and Its Impact on Single-Modal Judgement in Multimodal Perception
    Yuan, Tianyi
    Rau, Pei-Luen Patrick
    Zhao, Jingyu
    Zheng, Jian
    MULTISENSORY RESEARCH, 2023, 36 (05) : 387 - 411
  • [7] Multimodal Interaction Concepts for Mobile Augmented Reality Applications
    Hurst, Wolfgang
    van Wezel, Casper
    ADVANCES IN MULTIMEDIA MODELING, PT II, 2011, 6524 : 157 - 167
  • [8] Multimodal, Touchless Interaction in Spatial Augmented Reality Environments
    Elepfandt, Monika
    Suenderhauf, Marcelina
    DIGITAL HUMAN MODELING, 2011, 6777 : 263 - 271
  • [9] ARZombie: A Mobile Augmented Reality Game with Multimodal Interaction
    Cordeiro, Diogo
    Correia, Nuno
    Jesus, Rui
    PROCEEDINGS OF THE 2015 7TH INTERNATIONAL CONFERENCE ON INTELLIGENT TECHNOLOGIES FOR INTERACTIVE ENTERTAINMENT, 2015, : 22 - 31
  • [10] Multimodal Interaction Framework for Collaborative Augmented Reality in Education
    Asiri, Dalia Mohammed Eissa
    Allehaibi, Khalid Hamed
    Basori, Ahmad Hoirul
    INTERNATIONAL JOURNAL OF COMPUTER SCIENCE AND NETWORK SECURITY, 2022, 22 (07): : 268 - 282