Dual Track Multimodal Automatic Learning through Human-Robot Interaction

被引:0
|
作者
Jiang, Shuqiang [1 ,2 ]
Min, Weiqing [1 ,2 ]
Li, Xue [1 ,3 ]
Wang, Huayang [1 ,2 ]
Sun, Jian [1 ,3 ]
Zhou, Jiaqi [1 ]
机构
[1] Chinese Acad Sci, Key Lab Intelligent Informat Proc, Inst Comp Technol, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Beijing, Peoples R China
[3] Shandong Univ Sci & Technol, Qingdao, Peoples R China
基金
中国博士后科学基金; 北京市自然科学基金; 中国国家自然科学基金;
关键词
ONLINE;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Human beings are constantly improving their cognitive ability via automatic learning from the interaction with the environment. Two important aspects of automatic learning are the visual perception and knowledge acquisition. The fusion of these two aspects is vital for improving the intelligence and interaction performance of robots. Many automatic knowledge extraction and recognition methods have been widely studied. However, little work focuses on integrating automatic knowledge extraction and recognition into a unified framework to enable jointly visual perception and knowledge acquisition. To solve this problem, we propose a Dual Track Multimodal Automatic Learning (DTMAL) system, which consists of two components: Hybrid Incremental Learning (HIL) from the vision track and Multimodal Knowledge Extraction (MKE) from the knowledge track. HIL can incrementally improve recognition ability of the system by learning new object samples and new object concepts. MKE is capable of constructing and updating the multimodal knowledge items based on the recognized new objects from HIL and other knowledge by exploring the multimodal signals. The fusion of the two tracks is a mutual promotion process and jointly devote to the dual track learning. We have conducted the experiments through human-machine interaction and the experimental results validated the effectiveness of our proposed system.
引用
收藏
页码:4485 / 4491
页数:7
相关论文
共 50 条
  • [41] Learning representations for robust human-robot interaction
    Kuo, Yen-Ling
    AI MAGAZINE, 2024, 45 (04) : 561 - 568
  • [42] Comparing alternative modalities in the context of multimodal human-robot interaction
    Saren, Suprakas
    Mukhopadhyay, Abhishek
    Ghose, Debasish
    Biswas, Pradipta
    JOURNAL ON MULTIMODAL USER INTERFACES, 2024, 18 (01) : 69 - 85
  • [43] Learning cooperation from human-robot interaction
    Nicolescu, MN
    Mataric, MJ
    DISTRIBUTED AUTONOMOUS ROBOTIC SYSTEMS, 2000, : 477 - 478
  • [44] Learning Representations for Robust Human-Robot Interaction
    Kuo, Yen-Ling
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 20, 2024, : 22673 - 22673
  • [45] Research on multimodal human-robot interaction based on speech and gesture
    Deng Yongda
    Li Fang
    Xin Huang
    COMPUTERS & ELECTRICAL ENGINEERING, 2018, 72 : 443 - 454
  • [46] Multimodal emotion recognition with evolutionary computation for human-robot interaction
    Perez-Gaspar, Luis-Alberto
    Caballero-Morales, Santiago-Omar
    Trujillo-Romero, Felipe
    EXPERT SYSTEMS WITH APPLICATIONS, 2016, 66 : 42 - 61
  • [47] Speech to Head Gesture Mapping in Multimodal Human-Robot Interaction
    Aly, Amir
    Tapus, Adriana
    SERVICE ORIENTATION IN HOLONIC AND MULTI-AGENT MANUFACTURING CONTROL, 2012, 402 : 183 - 196
  • [48] Human-robot Interaction Control through Demonstration
    Lyu, Shangke
    Cheah, Chien Chern
    2018 26TH MEDITERRANEAN CONFERENCE ON CONTROL AND AUTOMATION (MED), 2018, : 1 - 6
  • [49] Learning and Comfort in Human-Robot Interaction: A Review
    Wang, Weitian
    Chen, Yi
    Li, Rui
    Jia, Yunyi
    APPLIED SCIENCES-BASEL, 2019, 9 (23):
  • [50] Incremental learning of gestures for human-robot interaction
    Okada, Shogo
    Kobayashi, Yoichi
    Ishibashi, Satoshi
    Nishida, Toyoaki
    AI & SOCIETY, 2010, 25 (02) : 155 - 168