Learn from Incomplete Tactile Data: Tactile Representation Learning with Masked Autoencoders

被引:1
|
作者
Cao, Guanqun [1 ]
Jiang, Jiaqi [2 ]
Bollegala, Danushka [1 ]
Luo, Shan [2 ]
机构
[1] Univ Liverpool, Dept Comp Sci, Liverpool L69 3BX, England
[2] Kings Coll London, Dept Engn, London WC2R 2LS, England
基金
英国工程与自然科学研究理事会;
关键词
OBJECT PROPERTIES; PERCEPTION;
D O I
10.1109/IROS55552.2023.10341788
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The missing signal caused by the objects being occluded or an unstable sensor is a common challenge during data collection. Such missing signals will adversely affect the results obtained from the data, and this issue is observed more frequently in robotic tactile perception. In tactile perception, due to the limited working space and the dynamic environment, the contact between the tactile sensor and the object is frequently insufficient and unstable, which causes the partial loss of signals, thus leading to incomplete tactile data. The tactile data will therefore contain fewer tactile cues with low information density. In this paper, we propose a tactile representation learning method, named TacMAE, based on Masked Autoencoder to address the problem of incomplete tactile data in tactile perception. In our framework, a portion of the tactile image is masked out to simulate the missing contact regions. By reconstructing the missing signals in the tactile image, the trained model can achieve a high-level understanding of surface geometry and tactile properties from limited tactile cues. The experimental results of tactile texture recognition show that TacMAE can achieve a high recognition accuracy of 71.4% in the zero-shot transfer and 85.8% after fine-tuning, which are 15.2% and 8.2% higher than the results without using masked modeling. The extensive experiments on YCB objects demonstrate the knowledge transferability of our proposed method and the potential to improve efficiency in tactile exploration.
引用
收藏
页码:10800 / 10805
页数:6
相关论文
共 50 条
  • [31] Shape recovery from passive locally dense tactile data
    Erdmann, M
    ROBOTICS: THE ALGORITHMIC PERSPECTIVE, 1998, : 119 - 132
  • [32] Tactile Map Automated Creation from Electronic Map Data
    Watanabe T.
    Kyokai Joho Imeji Zasshi/Journal of the Institute of Image Information and Television Engineers, 2019, 73 (06): : 1026 - 1029
  • [33] A training data augmentation approach using GAN for learning of tactile paving images
    Hayasaka, Hirofumi
    Kimura, Akio
    INTERNATIONAL WORKSHOP ON ADVANCED IMAGING TECHNOLOGY, IWAIT 2023, 2023, 12592
  • [34] An Application of Deep Learning to Tactile Data for Object Recognition under Visual Guidance
    Rouhafzay, Ghazal
    Cretu, Ana-Maria
    SENSORS, 2019, 19 (07):
  • [35] Real-Time Embedded Machine Learning for Tensorial Tactile Data Processing
    Ibrahim, Ali
    Valle, Maurizio
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2018, 65 (11) : 3897 - 3906
  • [36] Towards Developing an Algorithm in Discriminating Thin Material using Tri-axial Tactile Data from Human Tactile Sensation
    Mohammad Azzeim, Mat Jusoh
    Sugiman, Kenji
    Ohka, Masahiro
    Miyaoka, Tetsu
    Anitawati, Mohd Lokman
    2016 INTERNATIONAL SYMPOSIUM ON MICRO-NANOMECHATRONICS AND HUMAN SCIENCE (MHS), 2016,
  • [37] Learning to Live Life on the Edge: Online Learning for Data-Efficient Tactile Contour Following
    Stone, Elizabeth A.
    Lepora, Nathan F.
    Barton, David A. W.
    2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 9854 - 9860
  • [38] Walk and Learn: Facial Attribute Representation Learning from Egocentric Video and Contextual Data
    Wang, Jing
    Cheng, Yu
    Feris, Rogerio Schmidt
    2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 2295 - 2304
  • [39] SPARSE REPRESENTATION LEARNING OF DATA BY AUTOENCODERS WITH L1/2 REGULARIZATION
    Li, F.
    Zurada, J. M.
    Wu, W.
    NEURAL NETWORK WORLD, 2018, 28 (02) : 133 - 147
  • [40] Unsupervised Learning Enables Extraction of Tactile Information From Text Database
    Nagatomo, Tatsuho
    Hiraki, Takefumi
    Ishizuka, Hiroki
    Miki, Norihisa
    IEEE ACCESS, 2023, 11 (101155-101166): : 101155 - 101166