An efficient Multimodal language processor for parallel input strings in multimodal input fusion

被引:2
|
作者
Sun, Yong [1 ,2 ]
Shi, Yu [1 ]
Chen, Fang [1 ,2 ]
Chung, Vera [2 ]
机构
[1] Natl ICT Australia, Australian Technol Pk, Eveleigh, NSW 1430, Australia
[2] Univ Sydney, Sch IT, Sydney, NSW 2006, Australia
关键词
D O I
10.1109/ICSC.2007.61
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Multimodal User Interaction technology aims at building more natural and intuitive interfaces allowing a user to interact with a computer in a way similar to human-to-human communication, for example, through speech and gesture. As a critical component in Multimodal User Interaction, Multimodal Input Fusion explores the ways to effectively interpret the combined semantic interpretation of user inputs through multiple modalities. This paper proposes a new efficient unification-based multimodal language processor which can handle parallel input strings for Multimodal Input Fusion. With a structure sharing technology, it has the potential to achieve a low polynomial computational complexity while parsing multimodal inputs in versatile styles. The applicability of the proposed processor has been validated through an experiment with multimodal commands collected from traffic incident management scenarios. The description of the proposed multimodal language processor and preliminary experiment results are presented.
引用
收藏
页码:389 / +
页数:2
相关论文
共 50 条
  • [21] Software architecture for multimodal user input - FLUID
    Ilmonen, T
    Kontkanen, J
    UNIVERSAL ACCESS: THEORETICAL PERSPECTIVES, PRACTICE, AND EXPERIENCE, 2003, 2615 : 319 - 338
  • [22] Using Multimodal Input in Augmented Virtual Teleportation
    Sasikumar, Prasanth
    2022 IEEE CONFERENCE ON VIRTUAL REALITY AND 3D USER INTERFACES ABSTRACTS AND WORKSHOPS (VRW 2022), 2022, : 956 - 957
  • [23] Camouflaged target detection based on multimodal image input pixel-level fusion
    Peng, Ruihui
    Lai, Jie
    Yang, Xueting
    Sun, Dianxing
    Tan, Shuncheng
    Song, Yingjuan
    Guo, Wei
    FRONTIERS OF INFORMATION TECHNOLOGY & ELECTRONIC ENGINEERING, 2024, 25 (09) : 1226 - 1239
  • [24] UNO: Uncertainty-aware Noisy-Or Multimodal Fusion for Unanticipated Input Degradation
    Tian, Junjiao
    Cheung, Wesley
    Glaser, Nathaniel
    Liu, Yen-Cheng
    Kira, Zsolt
    2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 5716 - 5723
  • [25] Designing an Augmented Reality Multimodal Interface for 6DOF Manipulation Techniques Multimodal Fusion Using Gesture and Speech Input for AR
    Ismail, Ajune Wanis
    Billinghurst, Mark
    Sunar, Mohd Shahrizal
    Yusof, Cik Suhaimi
    INTELLIGENT SYSTEMS AND APPLICATIONS, VOL 1, 2019, 868 : 309 - 322
  • [26] UGaitNet: Multimodal Gait Recognition With Missing Input Modalities
    Marin-Jimenez, Manuel J.
    Castro, Francisco M.
    Delgado-Escano, Ruben
    Kalogeiton, Vicky
    Guil, Nicolas
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2021, 16 (16) : 5452 - 5462
  • [27] A usability study of multimodal input in an augmented reality environment
    Minkyung Lee
    Mark Billinghurst
    Woonhyuk Baek
    Richard Green
    Woontack Woo
    Virtual Reality, 2013, 17 : 293 - 305
  • [28] A usability study of multimodal input in an augmented reality environment
    Lee, Minkyung
    Billinghurst, Mark
    Baek, Woonhyuk
    Green, Richard
    Woo, Woontack
    VIRTUAL REALITY, 2013, 17 (04) : 293 - 305
  • [29] A Comparison of Input Entry Rates in a Multimodal Mobile Application
    Melto, Aleksi
    Turunen, Markku
    Hakulinen, Jaakko
    Kainulainen, Anssi
    Heimonen, Tomi
    INTERSPEECH 2008: 9TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2008, VOLS 1-5, 2008, : 2462 - 2465
  • [30] Antagonistic Video Generation Method Based on Multimodal Input
    Yu H.
    Yang X.
    Xu C.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2020, 57 (07): : 1522 - 1530