Multi-modal broad learning for material recognition

被引:3
|
作者
Wang, Zhaoxin [1 ,2 ]
Liu, Huaping [1 ,2 ]
Xu, Xinying [1 ,2 ]
Sun, Fuchun [1 ,2 ]
机构
[1] Tsinghua Univ, Dept Comp Sci & Technol, Beijing, Peoples R China
[2] Beijing Natl Res Ctr Informat Sci & Technol, State Key Lab Intelligent Technol & Syst, Beijing, Peoples R China
关键词
Human robot interaction - Learning systems;
D O I
10.1049/ccs2.12004
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Material recognition plays an important role in the interaction between robots and the external environment. For example, household service robots need to replace humans in the home environment to complete housework, so they need to interact with daily necessities and obtain their material performance. Images provide rich visual information about objects; however, it is often difficult to apply when objects are not visually distinct. In addition, tactile signals can be used to capture multiple characteristics of objects, such as texture, roughness, softness, and friction, which provides another crucial way for perception. How to effectively integrate multi-modal information is an urgent problem to be addressed. Therefore, a multi-modal material recognition framework CFBRL-KCCA for target recognition tasks is proposed in the paper. The preliminary features of each model are extracted by cascading broad learning, which is combined with the kernel canonical correlation learning, considering the differences among different models of heterogeneous data. Finally, the open dataset of household objects is evaluated. The results demonstrate that the proposed fusion algorithm provides an effective strategy for material recognition.
引用
收藏
页码:123 / 130
页数:8
相关论文
共 50 条
  • [31] Multi-modal recursive prompt learning with mixup embedding for generalization recognition
    Jia, Yunpeng
    Ye, Xiufen
    Liu, Yusong
    Guo, Shuxiang
    KNOWLEDGE-BASED SYSTEMS, 2024, 294
  • [32] Adaptive cross-fusion learning for multi-modal gesture recognition
    Benjia ZHOU
    Jun WAN
    Yanyan LIANG
    Guodong GUO
    虚拟现实与智能硬件(中英文), 2021, 3 (03) : 235 - 247
  • [33] Cybersecurity Named Entity Recognition Using Multi-Modal Ensemble Learning
    Yi, Feng
    Jiang, Bo
    Wang, Lu
    Wu, Jianjun
    IEEE ACCESS, 2020, 8 : 63214 - 63224
  • [34] Emotion recognition based on multi-modal physiological signals and transfer learning
    Fu, Zhongzheng
    Zhang, Boning
    He, Xinrun
    Li, Yixuan
    Wang, Haoyuan
    Huang, Jian
    FRONTIERS IN NEUROSCIENCE, 2022, 16
  • [35] Unequal adaptive visual recognition by learning from multi-modal data
    Cai, Ziyun
    Zhang, Tengfei
    Jing, Xiao-Yuan
    Shao, Ling
    INFORMATION SCIENCES, 2022, 600 : 1 - 21
  • [36] Multi-modal Broad Learning System for Medical Image and Text-based Classification
    Zhou, Yanhong
    Du, Jie
    Guan, Kai
    Wang, Tianfu
    2021 43RD ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE & BIOLOGY SOCIETY (EMBC), 2021, : 3439 - 3442
  • [37] Multi-Modal Multi-Action Video Recognition
    Shi, Zhensheng
    Liang, Ju
    Li, Qianqian
    Zheng, Haiyong
    Gu, Zhaorui
    Dong, Junyu
    Zheng, Bing
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 13658 - 13667
  • [38] Multi-modal and multi-granular learning
    Zhang, Bo
    Zhang, Ling
    ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PROCEEDINGS, 2007, 4426 : 9 - +
  • [39] Learning in an Inclusive Multi-Modal Environment
    Graham, Deryn
    Benest, Ian
    Nicholl, Peter
    JOURNAL OF CASES ON INFORMATION TECHNOLOGY, 2010, 12 (03) : 28 - 44
  • [40] Learning of Multi-Modal Stimuli in Hawkmoths
    Balkenius, Anna
    Dacke, Marie
    PLOS ONE, 2013, 8 (07):