Learning to Regrasp Using Visual-Tactile Representation-Based Reinforcement Learning

被引:0
|
作者
Zhang, Zhuangzhuang [1 ]
Sun, Han [1 ]
Zhou, Zhenning [1 ]
Wang, Yizhao [1 ]
Huang, Huang [2 ]
Zhang, Zhinan [1 ]
Cao, Qixin [1 ]
机构
[1] Shanghai Jiao Tong Univ, State Key Lab Mech Syst & Vibrat, Shanghai 200240, Peoples R China
[2] Beijing Inst Control Engn, Beijing 100191, Peoples R China
关键词
Visualization; Force; Grasping; Training; Representation learning; Tactile sensors; Feature extraction; Stability analysis; Optimization; Hardware; Reinforcement learning; representation learning; robotic regrasp; transfer learning; visual-tactile fusion; VISION; SENSOR;
D O I
10.1109/TIM.2024.3470030
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The open-loop grasp planner, which relies on vision, is prone to failure caused by calibration errors, visual occlusions, and other factors. Additionally, it cannot adapt the grasp pose and gripping force in real time, thereby increasing the risk of potential damage to unidentified objects. This work presents a multimodal regrasp control framework based on deep reinforcement learning (RL). Given a coarse initial grasp pose, the proposed regrasping policy efficiently optimizes grasp pose and gripping force by deeply fusing visual and high-resolution tactile data in a closed-loop fashion. To enhance the sample efficiency and generalization capability of the RL algorithm, this work leverages self-supervision to pretrain a visual-tactile representation model, which serves as a feature extraction network during RL policy training. The RL policy is trained purely in simulation and successfully deployed to a real-world environment via domain adaptation and domain randomization techniques. Extensive experimental results in simulation and real-world environments indicate that the robot guided by the regrasping policy is able to achieve gentle grasping of unknown objects with high success rates. Finally, the comparison results with the state-of-the-art algorithm also demonstrate the superiority of our algorithm.
引用
收藏
页数:11
相关论文
共 50 条
  • [11] Visual-Tactile Learning of Garment Unfolding for Robot-Assisted Dressing
    Zhang, Fan
    Demiris, Yiannis
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (09) : 5512 - 5519
  • [12] Visual-tactile learning of robotic cable-in-duct installation skills
    Duan, Boyi
    Qian, Kun
    Liu, Aohua
    Luo, Shan
    AUTOMATION IN CONSTRUCTION, 2025, 170
  • [13] Unidirectional Representation-Based Efficient Dictionary Learning
    Wang, Xiudong
    Li, Yali
    You, Shaodi
    Li, Hongdong
    Wang, Shengjin
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2020, 30 (01) : 59 - 74
  • [14] Visual-Tactile Robot Grasping Based on Human Skill Learning From Demonstrations Using a Wearable Parallel Hand Exoskeleton
    Lu, Zhenyu
    Chen, Lu
    Dai, Hengtai
    Li, Haoran
    Zhao, Zhou
    Zheng, Bofang
    Lepora, Nathan F. F.
    Yang, Chenguang
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (09) : 5384 - 5391
  • [15] Reinforcement-Learning Based Robotic Assembly of Fractured Objects Using Visual and Tactile Information
    Song, Xinchao
    Lamb, Nikolas
    Banerjee, Sean
    Banerjee, Natasha Kholgade
    2023 9TH INTERNATIONAL CONFERENCE ON AUTOMATION, ROBOTICS AND APPLICATIONS, ICARA, 2023, : 170 - 174
  • [16] Stable Reinforcement Learning with Autoencoders for Tactile and Visual Data
    van Hoof, Herke
    Chen, Nutan
    Kar, Maximilian
    van der Smagt, Patrick
    Peters, Jan
    2016 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS 2016), 2016, : 3928 - 3934
  • [17] Lifelong Visual-Tactile Cross-Modal Learning for Robotic Material Perception
    Zheng, Wendong
    Liu, Huaping
    Sun, Fuchun
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (03) : 1192 - 1203
  • [18] Environment Agnostic Representation for Visual Reinforcement learning
    Choi, Hyesong
    Lee, Hunsang
    Jeong, Seongwon
    Min, Dongbo
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 263 - 273
  • [19] Multiple Kernel Learning for Sparse Representation-Based Classification
    Shrivastava, Ashish
    Patel, Vishal M.
    Chellappa, Rama
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2014, 23 (07) : 3013 - 3024
  • [20] Constrained Visual Representation Learning With Bisimulation Metrics for Safe Reinforcement Learning
    Wang, Rongrong
    Cheng, Yuhu
    Wang, Xuesong
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2025, 34 : 379 - 393