Multi-modal broad learning for material recognition

被引:3
|
作者
Wang, Zhaoxin [1 ,2 ]
Liu, Huaping [1 ,2 ]
Xu, Xinying [1 ,2 ]
Sun, Fuchun [1 ,2 ]
机构
[1] Tsinghua Univ, Dept Comp Sci & Technol, Beijing, Peoples R China
[2] Beijing Natl Res Ctr Informat Sci & Technol, State Key Lab Intelligent Technol & Syst, Beijing, Peoples R China
关键词
Human robot interaction - Learning systems;
D O I
10.1049/ccs2.12004
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Material recognition plays an important role in the interaction between robots and the external environment. For example, household service robots need to replace humans in the home environment to complete housework, so they need to interact with daily necessities and obtain their material performance. Images provide rich visual information about objects; however, it is often difficult to apply when objects are not visually distinct. In addition, tactile signals can be used to capture multiple characteristics of objects, such as texture, roughness, softness, and friction, which provides another crucial way for perception. How to effectively integrate multi-modal information is an urgent problem to be addressed. Therefore, a multi-modal material recognition framework CFBRL-KCCA for target recognition tasks is proposed in the paper. The preliminary features of each model are extracted by cascading broad learning, which is combined with the kernel canonical correlation learning, considering the differences among different models of heterogeneous data. Finally, the open dataset of household objects is evaluated. The results demonstrate that the proposed fusion algorithm provides an effective strategy for material recognition.
引用
收藏
页码:123 / 130
页数:8
相关论文
共 50 条
  • [41] Reliable Multi-modal Learning: A Survey
    Yang Y.
    Zhan D.-C.
    Jiang Y.
    Xiong H.
    Ruan Jian Xue Bao/Journal of Software, 2021, 32 (04): : 1067 - 1081
  • [42] Multi-Modal Meta Continual Learning
    Gai, Sibo
    Chen, Zhengyu
    Wang, Donglin
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [43] Learning multi-modal control programs
    Mehta, TR
    Egerstedt, M
    HYBRID SYSTEMS: COMPUTATION AND CONTROL, 2005, 3414 : 466 - 479
  • [44] Imagery in multi-modal object learning
    Jüttner, M
    Rentschler, I
    BEHAVIORAL AND BRAIN SCIENCES, 2002, 25 (02) : 197 - +
  • [45] Multi-modal Network Representation Learning
    Zhang, Chuxu
    Jiang, Meng
    Zhang, Xiangliang
    Ye, Yanfang
    Chawla, Nitesh, V
    KDD '20: PROCEEDINGS OF THE 26TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2020, : 3557 - 3558
  • [46] Modelling multi-modal learning in a hawkmoth
    Balkenius, Anna
    Kelber, Almut
    Balkenius, Christian
    FROM ANIMALS TO ANIMATS 9, PROCEEDINGS, 2006, 4095 : 422 - 433
  • [47] On Multi-modal Fusion for Freehand Gesture Recognition
    Schak, Monika
    Gepperth, Alexander
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2020, PT I, 2020, 12396 : 862 - 873
  • [48] Multi-modal Attention for Speech Emotion Recognition
    Pan, Zexu
    Luo, Zhaojie
    Yang, Jichen
    Li, Haizhou
    INTERSPEECH 2020, 2020, : 364 - 368
  • [49] Multi-Modal Multiple-Instance Learning with the Application to the Cannabis Webpage Recognition
    Wang, Yinjuan
    Xie, Nianhua
    Hu, Weiming
    Yang, Jinfeng
    2011 FIRST ASIAN CONFERENCE ON PATTERN RECOGNITION (ACPR), 2011, : 105 - 109
  • [50] Electromagnetic signal feature fusion and recognition based on multi-modal deep learning
    Hou C.
    Zhang X.
    Chen X.
    International Journal of Performability Engineering, 2020, 16 (06): : 941 - 949