An efficient Multimodal language processor for parallel input strings in multimodal input fusion

被引:2
|
作者
Sun, Yong [1 ,2 ]
Shi, Yu [1 ]
Chen, Fang [1 ,2 ]
Chung, Vera [2 ]
机构
[1] Natl ICT Australia, Australian Technol Pk, Eveleigh, NSW 1430, Australia
[2] Univ Sydney, Sch IT, Sydney, NSW 2006, Australia
关键词
D O I
10.1109/ICSC.2007.61
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Multimodal User Interaction technology aims at building more natural and intuitive interfaces allowing a user to interact with a computer in a way similar to human-to-human communication, for example, through speech and gesture. As a critical component in Multimodal User Interaction, Multimodal Input Fusion explores the ways to effectively interpret the combined semantic interpretation of user inputs through multiple modalities. This paper proposes a new efficient unification-based multimodal language processor which can handle parallel input strings for Multimodal Input Fusion. With a structure sharing technology, it has the potential to achieve a low polynomial computational complexity while parsing multimodal inputs in versatile styles. The applicability of the proposed processor has been validated through an experiment with multimodal commands collected from traffic incident management scenarios. The description of the proposed multimodal language processor and preliminary experiment results are presented.
引用
收藏
页码:389 / +
页数:2
相关论文
共 50 条
  • [1] Gestures and multimodal input
    Keates, S
    Robinson, P
    BEHAVIOUR & INFORMATION TECHNOLOGY, 1999, 18 (01) : 36 - 44
  • [2] Multimodal input in second-language speech processing
    Hardison, Debra M.
    LANGUAGE TEACHING, 2021, 54 (02) : 206 - 220
  • [3] Language Acquisition: The Emergence of Words from Multimodal Input
    ten Bosch, Louis
    Boves, Lou
    TEXT, SPEECH AND DIALOGUE, PROCEEDINGS, 2008, 5246 : 261 - 268
  • [4] Multimodal Fusion: Gesture and Speech Input in Augmented Reality Environment
    Ismail, Ajune Wanis
    Sunar, Mohd Shahrizal
    COMPUTATIONAL INTELLIGENCE IN INFORMATION SYSTEMS, 2015, 331 : 245 - 254
  • [5] MIMOQA: Multimodal Input Multimodal Output Question Answering
    Singh, Hrituraj
    Nasery, Anshul
    Mehta, Denil
    Agarwal, Aishwarya
    Lamba, Jatin
    Srinivasan, Balaji Vasan
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 5317 - 5332
  • [6] INTRODUCTION TO THE SPECIAL ISSUE LANGUAGE LEARNING FROM MULTIMODAL INPUT
    Peters, Elke
    Munoz, Carmen
    STUDIES IN SECOND LANGUAGE ACQUISITION, 2020, 42 (03) : 489 - 497
  • [7] Teachers' voices on multimodal input for second or foreign language learning
    Cardenas-Claros, Monica Stella
    Sydorenko, Tetyana
    Huntley, Elizabeth
    Perez, Maribel Montero
    LANGUAGE TEACHING RESEARCH, 2023,
  • [8] MULTIMODAL INPUT IN SLA RESEARCH
    Perez, Maribel Montero
    STUDIES IN SECOND LANGUAGE ACQUISITION, 2020, 42 (03) : 653 - 663
  • [9] Input modelling for multimodal data
    Cheng, Russell C. H.
    Currie, Christine S. M.
    JOURNAL OF THE OPERATIONAL RESEARCH SOCIETY, 2020, 71 (06) : 1038 - 1052
  • [10] A new approach for integrating multimodal input via late semantic fusion
    McGlaun, G.
    Althoff, F.
    Lang, M.
    VDI Berichte, 2002, (1678): : 181 - 186