Visual Spatial Attention and Proprioceptive Data-Driven Reinforcement Learning for Robust Peg-in-Hole Task Under Variable Conditions

被引:12
|
作者
Yasutomi, Andre Yuji [1 ]
Ichiwara, Hideyuki [1 ]
Ito, Hiroshi [1 ]
Mori, Hiroki [2 ]
Ogata, Tetsuya [3 ,4 ]
机构
[1] Hitachi Ltd, R&D Grp, Hitachinaka 3120034, Japan
[2] Waseda Univ, Future Robot Org, Tokyo 1698555, Japan
[3] Waseda Univ, Grad Sch Fundamental Sci & Engn, Tokyo 1698555, Japan
[4] Waseda Univ, Waseda Res Inst Sci & Engn WISE, Tokyo 1698555, Japan
关键词
Robotics and automation in construction; reinforcement learning; deep learning for visual perception;
D O I
10.1109/LRA.2023.3243526
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Anchor-bolt insertion is a peg-in-hole task performed in the construction field for holes in concrete. Efforts have been made to automate this task, but the variable lighting and hole surface conditions, as well as the requirements for short setup and task execution time make the automation challenging. In this study, we introduce a vision and proprioceptive data-driven robot control model for this task that is robust to challenging lighting and hole surface conditions. This model consists of a spatial attention point network (SAP) and a deep reinforcement learning (DRL) policy that are trained jointly end-to-end to control the robot. The model is trained in an offline manner, with a sample-efficient framework designed to reduce training time and minimize the reality gap when transferring the model to the physical world. Through evaluations with an industrial robot performing the task in 12 unknown holes, starting from 16 different initial positions, and under three different lighting conditions (two with misleading shadows), we demonstrate that SAP can generate relevant attention points of the image even in challenging lighting conditions. We also show that the proposed model enables task execution with higher success rate and shorter task completion time than various baselines. Due to the proposed model's high effectiveness even in severe lighting, initial positions, and hole conditions, and the offline training framework's high sample-efficiency and short training time, this approach can be easily applied to construction.
引用
收藏
页码:1834 / 1841
页数:8
相关论文
共 41 条
  • [1] Data-driven Object-centric Framework for Peg-in-hole Learning
    Lee G.
    Lee J.
    Noh S.
    Bak S.
    Lee K.
    Journal of Institute of Control, Robotics and Systems, 2023, 29 (04) : 355 - 359
  • [2] Deep Reinforcement Learning for Peg-in-hole Assembly Task Via Information Utilization Method
    Wang, Fei
    Cui, Ben
    Liu, Yue
    Ren, Baiming
    JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 2022, 106 (01)
  • [3] Deep Reinforcement Learning for Peg-in-hole Assembly Task Via Information Utilization Method
    Fei Wang
    Ben Cui
    Yue Liu
    Baiming Ren
    Journal of Intelligent & Robotic Systems, 2022, 106
  • [4] Variable Compliance Control for Robotic Peg-in-Hole Assembly: A Deep-Reinforcement-Learning Approach
    Beltran-Hernandez, Cristian C.
    Petit, Damien
    Ramirez-Alpizar, Ixchel G.
    Harada, Kensuke
    APPLIED SCIENCES-BASEL, 2020, 10 (19): : 1 - 17
  • [5] Data-Driven Robust Control Using Reinforcement Learning
    Ngo, Phuong D.
    Tejedor, Miguel
    Godtliebsen, Fred
    APPLIED SCIENCES-BASEL, 2022, 12 (04):
  • [6] Peg-in-hole assembly skill imitation learning method based on ProMPs under task geometric representation
    Zang, Yajing
    Wang, Pengfei
    Zha, Fusheng
    Guo, Wei
    Zheng, Chao
    Sun, Lining
    FRONTIERS IN NEUROROBOTICS, 2023, 17
  • [7] Fuzzy Logic-Driven Variable Time-Scale Prediction-Based Reinforcement Learning for Robotic Multiple Peg-in-Hole Assembly
    Hou, Zhimin
    Li, Zhihu
    Hsu, Chenwei
    Zhang, Kuangen
    Xu, Jing
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2022, 19 (01) : 218 - 229
  • [8] DATA-DRIVEN ROBUST MULTI-AGENT REINFORCEMENT LEARNING
    Wang, Yudan
    Wang, Yue
    Zhou, Yi
    Velasquez, Alvaro
    Zou, Shaofeng
    2022 IEEE 32ND INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING (MLSP), 2022,
  • [9] VRL3: A Data-Driven Framework for Visual Deep Reinforcement Learning
    Wang, Che
    Luo, Xufang
    Ross, Keith
    Li, Dongsheng
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [10] Data-Driven Aircraft Modeling for Robust Reinforcement Learning Control Synthesis With Flight Test Validation
    Benyamen, Hady
    Chowdhury, Mozammal
    Keshmiri, Shawn
    JOURNAL OF DYNAMIC SYSTEMS MEASUREMENT AND CONTROL-TRANSACTIONS OF THE ASME, 2024, 146 (06):