Adaptive disassembly sequence planning for VR maintenance training via deep reinforcement learning

被引:21
|
作者
Mao, Haoyang [1 ]
Liu, Zhenyu [1 ]
Qiu, Chan [1 ]
机构
[1] Zhejiang Univ, State Key Lab CAD&CG, Hangzhou 310027, Peoples R China
基金
中国国家自然科学基金;
关键词
Disassembly sequence planning; Deep reinforcement learning; Genetic algorithm; VR maintenance training; VIRTUAL-REALITY; ALGORITHM; INTELLIGENCE; SYSTEM;
D O I
10.1007/s00170-021-08290-x
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
VR training equipped with meta-heuristic disassembly planning algorithms has been widely applied in pre-employment training in recent years. However, these algorithms are usually authored for specific sequences of a single product, and it remains a challenge to generalize them to maintenance training with unpredictable disassembly targets. As a promising method for settling dynamic and stochastic problems, deep reinforcement learning (DRL) provides a new insight to dynamically generate optimal sequences. This study introduces the deep Q-network (DQN), a successful DRL method, to fulfill adaptive disassembly sequence planning (DSP) for the VR maintenance training. Disassembly Petri net is established to describe the disassembly process, and then the DSP problem is defined as a Markov decision process that can be solved by DQN. Two neural networks are designed and updated asynchronously, and the training of DQN is further achieved by backpropagation of errors. Especially, we replace the long-term return in DQN with the fitness function of the genetic algorithm to avoid dependence on the immediate reward. Several experiments have been carried out to exhibit great potentials of our method in on-site maintenance where the fault is uncertain.
引用
收藏
页码:3039 / 3048
页数:10
相关论文
共 50 条
  • [21] Spatial planning of urban communities via deep reinforcement learning
    Yu Zheng
    Yuming Lin
    Liang Zhao
    Tinghai Wu
    Depeng Jin
    Yong Li
    Nature Computational Science, 2023, 3 : 748 - 762
  • [22] Spatial planning of urban communities via deep reinforcement learning
    Zheng, Yu
    Lin, Yuming
    Zhao, Liang
    Wu, Tinghai
    Jin, Depeng
    Li, Yong
    NATURE COMPUTATIONAL SCIENCE, 2023, 3 (09): : 748 - +
  • [23] Network Maintenance Planning Via Multi-Agent Reinforcement Learning
    Thomas, Jonathan
    Hernandez, Marco Perez
    Parlikad, Ajith Kumar
    Piechocki, Robert
    2021 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2021, : 2289 - 2295
  • [24] Research on Virtual Maintenance Disassembly Sequence Planning for Weapons Equipment
    Ye Lin
    Shen Yan-an
    CONFERENCE PROCEEDINGS OF 2017 3RD IEEE INTERNATIONAL CONFERENCE ON CONTROL SCIENCE AND SYSTEMS ENGINEERING (ICCSSE), 2017, : 706 - 710
  • [25] Training in VR: A Preliminary Study on Learning Assembly/Disassembly Sequences
    Sportillo, Daniele
    Avveduto, Giovanni
    Tecchia, Franco
    Carrozzino, Marcello
    AUGMENTED AND VIRTUAL REALITY, AVR 2015, 2015, 9254 : 332 - 343
  • [26] Efficient Deep Reinforcement Learning via Adaptive Policy Transfer
    Yang, Tianpei
    Hao, Jianye
    Meng, Zhaopeng
    Zhang, Zongzhang
    Hu, Yujing
    Chen, Yingfeng
    Fan, Changjie
    Wang, Weixun
    Liu, Wulong
    Wang, Zhaodong
    Peng, Jiajie
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 3094 - 3100
  • [27] Deep Reinforcement Learning With Dynamic Graphs for Adaptive Informative Path Planning
    Vashisth, Apoorva
    Rueckin, Julius
    Magistri, Federico
    Stachniss, Cyrill
    Popovic, Marija
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (09): : 7747 - 7754
  • [28] Adaptive speed planning for Unmanned Vehicle Based on Deep Reinforcement Learning
    Liu, Hao
    Shen, Yi
    Zhou, Wenjing
    Zou, Yuelin
    Zhou, Chang
    He, Shuyao
    2024 5TH INTERNATIONAL CONFERENCE ON MECHATRONICS TECHNOLOGY AND INTELLIGENT MANUFACTURING, ICMTIM 2024, 2024, : 642 - 645
  • [29] A deep reinforcement learning assisted simulated annealing algorithm for a maintenance planning problem
    Kosanoglu, Fuat
    Atmis, Mahir
    Turan, Hasan Huseyin
    ANNALS OF OPERATIONS RESEARCH, 2024, 339 (1-2) : 79 - 110
  • [30] Learning to Adapt: Communication Load Balancing via Adaptive Deep Reinforcement Learning
    Wu, Di
    Xu, Yi Tian
    Li, Jimmy
    Jenkin, Michael
    Hossain, Ekram
    Jang, Seowoo
    Xin, Yan
    Zhang, Charlie
    Liu, Xue
    Dudek, Gregory
    IEEE CONFERENCE ON GLOBAL COMMUNICATIONS, GLOBECOM, 2023, : 2973 - 2978