Reinforcement Learning Enabled Self-Homing of Industrial Robotic Manipulators in Manufacturing

被引:0
|
作者
Karigiannis, John N. [1 ]
Laurin, Philippe [3 ]
Liu, Shaopeng [1 ]
Holovashchenko, Viktor [1 ]
Lizotte, Antoine [2 ]
Roux, Vincent [2 ]
Boulet, Philippe [2 ]
机构
[1] GE Res, 1 Res Circle, Niskayuna, NY 12309 USA
[2] Global Robot & Automat Ctr GE Aviat, 2 Blvd Aeroport, Bromont, PQ J2L 1A3, Canada
[3] Robotech Automatisat, 2168 Rue Prov, Longueuil, PQ J4G 1R7, Canada
关键词
reinforcement learning; self-homing; parallel-agent; industrial robotic manipulator;
D O I
暂无
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
Industrial robotics has been playing a major role in manufacturing across all types of industries. One common task of robotic cells in manufacturing is called homing, a step that enables a robotic arm to return to its initial / home position (HPos) from anywhere in a robotic cell, without collision or experiencing robot singularities while respecting its joint limits. In almost all industrial robotic cells, an operation cycle starts from, and ends to HPos. The home position also works as a safe state for a cycle to restart when an alarm or fault occurs within the cell. When an alarm occurs, the robot configuration in the cell is unpredictable, thus challenging to bring the robot, autonomously and with safety at HPos and restart the operation. This paper presents a non-vision, reinforcement learning-based approach of a parallel-agent setting to enable selfhoming capability in industrial robotic cells, eliminating the need of manual programming of robot manipulators. This approach assumes the sensing of an unknown robotic cell environment pre-encoded in the state definition so that the policies learned can be transferred without further training. The agents are trained in a simulation environment generated by the mechanical design of an actual robotic cell to increase the accuracy of mapping the real environment to the simulated one. The approach explores the impact of certain curriculum on the agent's learning and evaluates two choices, compared to a non-curriculum baseline. A parallel-agent, multi-process training setting is employed to enhance performance in exploring the state space, where experiences are shared among the agents via shared memory. Upon deployment, all agents are involved with their respective policies in a collective manner. The approach has been demonstrated in simulated industrial robotic cells, and it has been shown that the policies derived in simulation are transferable to a corresponding real industrial robotic cells, and are generalizable to other robotic systems in manufacturing settings. (C) 2022 Society of Manufacturing Engineers (SME). Published by Elsevier Ltd. All rights reserved. Peer-review under responsibility of the Scientific Committee of the NAMRI/SME.
引用
收藏
页码:909 / 918
页数:10
相关论文
共 50 条
  • [31] Blockchain-Enabled Data Collection and Sharing for Industrial IoT With Deep Reinforcement Learning
    Liu, Chi Harold
    Lin, Qiuxia
    Wen, Shilin
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2019, 15 (06) : 3516 - 3526
  • [32] Deep Reinforcement Learning Based Computation Offloading in Fog Enabled Industrial Internet of Things
    Ren, Yijing
    Sun, Yaohua
    Peng, Mugen
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2021, 17 (07) : 4978 - 4987
  • [33] Fuzzy-logic-based reinforcement learning of admittance control for automated robotic manufacturing
    Prabhu, SM
    Garg, DP
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 1998, 11 (01) : 7 - 23
  • [34] Robotic disassembly of screws for end-of-life product remanufacturing enabled by deep reinforcement learning
    Peng, Yiqun
    Li, Weidong
    Liang, Yuchen
    Pham, Duc Truong
    JOURNAL OF CLEANER PRODUCTION, 2024, 439
  • [35] Self-repair of smart manufacturing systems by deep reinforcement learning
    Epureanu, Bogdan, I
    Li, Xingyu
    Nassehi, Aydin
    Koren, Yoram
    CIRP ANNALS-MANUFACTURING TECHNOLOGY, 2020, 69 (01) : 421 - 424
  • [36] Visualizing Multi-Agent Reinforcement Learning for Robotic Communication in Industrial IoT Networks
    Luo, Ruyu
    Ni, Wanli
    Tian, Hui
    IEEE INFOCOM 2022 - IEEE CONFERENCE ON COMPUTER COMMUNICATIONS WORKSHOPS (INFOCOM WKSHPS), 2022,
  • [37] Out-of-order execution enabled deep reinforcement learning for dynamic additive manufacturing scheduling
    Sun, Mingyue
    Ding, Jiyuchen
    Zhao, Zhiheng
    Chen, Jian
    Huang, George Q.
    Wang, Lihui
    ROBOTICS AND COMPUTER-INTEGRATED MANUFACTURING, 2025, 91
  • [38] Fault self-healing: A biological immune heuristic reinforcement learning method with root cause reasoning in industrial manufacturing process
    Tian, JiaYi
    Yin, Ming
    Jiang, Jijiao
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 133
  • [39] Path planning of 6-DOF free-floating space robotic manipulators using reinforcement learning
    Al Ali, Ahmad
    Shi, Jian-Feng
    Zhu, Zheng H.
    ACTA ASTRONAUTICA, 2024, 224 : 367 - 378
  • [40] Preliminary Study of Self-adaptive Constrained Output Iterative Learning Control for Robotic Manipulators
    Yovchev, Kaloyan
    Miteva, Lyubomira
    ADVANCES IN SERVICE AND INDUSTRIAL ROBOTICS, RAAD 2021, 2021, 102 : 275 - 283