Multimodal fusion recognition for digital twin

被引:4
|
作者
Zhou, Tianzhe [1 ]
Zhang, Xuguang [1 ]
Kang, Bing [1 ]
Chen, Mingkai [1 ]
机构
[1] Nanjing Univ Posts & Telecommun, Key Lab Broadband Wireless Commun & Sensor Network, Minist Educ, Nanjing 210003, Peoples R China
关键词
Digital twin; Multimodal fusion; Object recognition; Deep learning; Transfer learning; CLASSIFICATION; NETWORKS; FEATURES;
D O I
10.1016/j.dcan.2022.10.009
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
The digital twin is the concept of transcending reality, which is the reverse feedback from the real physical space to the virtual digital space. People hold great prospects for this emerging technology. In order to realize the upgrading of the digital twin industrial chain, it is urgent to introduce more modalities, such as vision, haptics, hearing and smell, into the virtual digital space, which assists physical entities and virtual objects in creating a closer connection. Therefore, perceptual understanding and object recognition have become an urgent hot topic in the digital twin. Existing surface material classification schemes often achieve recognition through machine learning or deep learning in a single modality, ignoring the complementarity between multiple modalities. In order to overcome this dilemma, we propose a multimodal fusion network in our article that combines two modalities, visual and haptic, for surface material recognition. On the one hand, the network makes full use of the potential correlations between multiple modalities to deeply mine the modal semantics and complete the data mapping. On the other hand, the network is extensible and can be used as a universal architecture to include more modalities. Experiments show that the constructed multimodal fusion network can achieve 99.42% classification accuracy while reducing complexity.
引用
收藏
页码:337 / 346
页数:10
相关论文
共 50 条
  • [21] Multimodal data visualization method for digital twin campus construction
    Xie, Yakun
    Zhan, Ni
    Zhu, Qing
    Zhan, Junjie
    Guo, Zhihao
    Qiao, Changjian
    Zhu, Jun
    Xu, Bingli
    INTERNATIONAL JOURNAL OF DIGITAL EARTH, 2024, 17 (01)
  • [22] Digital Twin Heuristic Positioning of Insulation in Multimodal Electric Systems
    Sikora, Andrzej
    Zielonka, Adam
    Ijaz, Muhammad Fazal
    Wozniak, Marcin
    IEEE TRANSACTIONS ON CONSUMER ELECTRONICS, 2024, 70 (01) : 3436 - 3445
  • [23] New method for modeling digital twin behavior perception of cows: Cow daily behavior recognition based on multimodal data
    Zhang, Yi
    Zhang, Yu
    Jiang, Hailong
    Du, Haitao
    Xue, Aidi
    Shen, Weizheng
    COMPUTERS AND ELECTRONICS IN AGRICULTURE, 2024, 226
  • [24] Multimodal Data Fusion Architectures in Audiovisual Speech Recognition
    Sayed, Hadeer M.
    ElDeeb, Hesham E.
    Taiel, Shereen A.
    INFORMATION SYSTEMS AND TECHNOLOGIES, VOL 2, WORLDCIST 2023, 2024, 800 : 655 - 667
  • [25] Multimodal features fusion for gait, gender and shoes recognition
    Castro, Francisco M.
    Marin-Jimenez, Manuel J.
    Guil, Nicolas
    MACHINE VISION AND APPLICATIONS, 2016, 27 (08) : 1213 - 1228
  • [26] A Framework to Evaluate Fusion Methods for Multimodal Emotion Recognition
    Pena, Diego
    Aguilera, Ana
    Dongo, Irvin
    Heredia, Juanpablo
    Cardinale, Yudith
    IEEE ACCESS, 2023, 11 : 10218 - 10237
  • [27] EXPLOITING MULTIMODAL DATA FUSION IN ROBUST SPEECH RECOGNITION
    Heracleous, Panikos
    Badin, Pierre
    Bailly, Gerard
    Hagita, Norihiro
    2010 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME 2010), 2010, : 568 - 572
  • [28] Multimodal recognition based on fusion of ear and profile face
    Xu, Xiaona
    Mu, Zhichun
    PROCEEDINGS OF THE FOURTH INTERNATIONAL CONFERENCE ON IMAGE AND GRAPHICS, 2007, : 598 - +
  • [29] Finger Multimodal Features Fusion and Recognition Based on CNN
    Wang, Li
    Zhang, Haigang
    Yang, Jingfeng
    2019 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI 2019), 2019, : 3183 - 3188
  • [30] Dual Memory Fusion for Multimodal Speech Emotion Recognition
    Priyasad, Darshana
    Fernando, Tharindu
    Sridharan, Sridha
    Denman, Simon
    Fookes, Clinton
    INTERSPEECH 2023, 2023, : 4543 - 4547