Deep reinforcement learning based proactive dynamic obstacle avoidance for safe human-robot collaboration

被引:0
|
作者
Xia, Wanqing [1 ]
Lu, Yuqian [1 ]
Xu, Weiliang [1 ]
Xu, Xun [1 ]
机构
[1] Univ Auckland, 20 Symond St, Auckland 1010, New Zealand
关键词
Human-robot collaboration; Dynamic obstacle avoidance; Deep reinforcement learning; Reward engineering;
D O I
10.1016/j.mfglet.2024.09.151
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
Ensuring the health and safety of human operators is paramount in manufacturing, particularly in human-robot collaborative environments. In this paper, we present a deep reinforcement learning-based trajectory planning method for a robotic manipulator designed to avoid collisions with human body parts in real-time while achieving its goal. We modelled the human arm as a freely moving cylinder in 3D space and formulated the dynamic obstacle avoidance problem as a Markov decision process. The algorithm was tested in a simulated environment that closely mimics our laboratory environment, with the goal of training a deep reinforcement learning model for autonomous task completion. A composite reward function was developed to balance the effects of different environmental variables, and the soft-actor critic algorithm was employed. The trained model demonstrated a 93% success rate in avoiding dynamic obstacles while achieving its goals when tested on a generated data set. (c) 2024 The Authors. Published by ELSEVIER Ltd. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/)
引用
收藏
页码:1246 / 1256
页数:11
相关论文
共 50 条
  • [21] Mobile robot dynamic obstacle avoidance method based on improved reinforcement learning
    Xu J.
    Shao K.
    Wang J.
    Liu X.
    Zhongguo Guanxing Jishu Xuebao/Journal of Chinese Inertial Technology, 2023, 31 (01): : 92 - 99
  • [22] A Dynamic Planner for Safe and Predictable Human-Robot Collaboration
    Pupa, Andrea
    Minelli, Marco
    Secchi, Cristian
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (01) : 507 - 514
  • [23] A Framework and Algorithm for Human-Robot Collaboration Based on Multimodal Reinforcement Learning
    Cai, Zeyuan
    Feng, Zhiquan
    Zhou, Liran
    Ai, Changsheng
    Shao, Haiyan
    Yang, Xiaohui
    COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2022, 2022
  • [24] Learning Controllers for Reactive and Proactive Behaviors in Human-Robot Collaboration
    Rozo, Leonel
    Silverio, Joao
    Calinon, Sylvain
    Caldwell, Darwin G.
    FRONTIERS IN ROBOTICS AND AI, 2016, 3
  • [25] Dynamic Obstacle Avoidance Method for Carrier Aircraft Based on Deep Reinforcement Learning
    Xue J.
    Kong X.
    Guo Y.
    Lu A.
    Li J.
    Wan X.
    Xu M.
    Jisuanji Fuzhu Sheji Yu Tuxingxue Xuebao/Journal of Computer-Aided Design and Computer Graphics, 2021, 33 (07): : 1102 - 1112
  • [26] Obstacle Avoidance Planning of Virtual Robot Picking Path Based on Deep Reinforcement Learning
    Xiong J.
    Li Z.
    Chen S.
    Zheng Z.
    Nongye Jixie Xuebao/Transactions of the Chinese Society for Agricultural Machinery, 2020, 51 : 1 - 10
  • [27] Deep Reinforcement Learning of Map-Based Obstacle Avoidance for Mobile Robot Navigation
    Chen G.
    Pan L.
    Chen Y.
    Xu P.
    Wang Z.
    Wu P.
    Ji J.
    Chen X.
    SN Computer Science, 2021, 2 (6)
  • [28] Dynamic Obstacle Avoidance for Application of Human-Robot Cooperative Dispensing Medicines
    Wang Z.
    Xu H.
    Lü N.
    Tao W.
    Chen G.
    Chi W.
    Sun L.
    Journal of Shanghai Jiaotong University (Science), 2022, 27 (01): : 24 - 35
  • [29] Research on avoidance of dynamic obstacle and singularity for manipulator in human-robot cooperation
    Cao Q.
    Sun M.
    Xue W.
    Xia S.
    Huazhong Keji Daxue Xuebao (Ziran Kexue Ban)/Journal of Huazhong University of Science and Technology (Natural Science Edition), 2020, 48 (01): : 55 - 59and65
  • [30] Reinforcement Learning for Mobile Robot Obstacle Avoidance Under Dynamic Environments
    Huang, Liwei
    Qu, Hong
    Fu, Mingsheng
    Deng, Wu
    PRICAI 2018: TRENDS IN ARTIFICIAL INTELLIGENCE, PT I, 2018, 11012 : 441 - 453