Actor-Critic Deep Reinforcement Learning for Solving Job Shop Scheduling Problems

被引:142
|
作者
Liu, Chien-Liang [1 ]
Chang, Chuan-Chin [1 ]
Tseng, Chun-Jan [1 ]
机构
[1] Natl Chiao Tung Univ, Dept Ind Engn & Management, Hsinchu 30010, Taiwan
来源
IEEE ACCESS | 2020年 / 8卷
关键词
Job shop scheduling; Machine learning; Benchmark testing; Dynamic scheduling; Learning (artificial intelligence); Training; Optimization; Job shop scheduling problem ([!text type='JS']JS[!/text]SP); deep reinforcement learning; actor-critic network; parallel training; OPTIMIZATION; SEARCH; LEVEL; GAME; GO;
D O I
10.1109/ACCESS.2020.2987820
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In the past decades, many optimization methods have been devised and applied to job shop scheduling problem (JSSP) to find the optimal solution. Many methods assumed that the scheduling results were applied to static environments, but the whole environments in the real world are always dynamic. Moreover, many unexpected events such as machine breakdowns and material problems may be present to adversely affect the initial job scheduling. This work views JSSP as a sequential decision making problem and proposes to use deep reinforcement learning to cope with this problem. The combination of deep learning and reinforcement learning avoids handcraft features as used in traditional reinforcement learning, and it is expected that the combination will make the whole learning phase more efficient. Our proposed model comprises actor network and critic network, both including convolution layers and fully connected layer. Actor network agent learns how to behave in different situations, while critic network helps agent evaluate the value of statement then return to actor network. This work proposes a parallel training method, combining asynchronous update as well as deep deterministic policy gradient (DDPG), to train the model. The whole network is trained with parallel training on a multi-agent environment and different simple dispatching rules are considered as actions. We evaluate our proposed model on more than ten instances that are present in a famous benchmark problem library - OR library. The evaluation results indicate that our method is comparative in static JSSP benchmark problems, and achieves a good balance between makespan and execution time in dynamic environments. Scheduling score of our method is 91.12% in static JSSP benchmark problems, and 80.78% in dynamic environments.
引用
收藏
页码:71752 / 71762
页数:11
相关论文
共 50 条
  • [21] A Deep Actor-Critic Reinforcement Learning Framework for Dynamic Multichannel Access
    Zhong, Chen
    Lu, Ziyang
    Gursoy, M. Cenk
    Velipasalar, Senem
    IEEE TRANSACTIONS ON COGNITIVE COMMUNICATIONS AND NETWORKING, 2019, 5 (04) : 1125 - 1139
  • [22] A World Model for Actor-Critic in Reinforcement Learning
    Panov, A. I.
    Ugadiarov, L. A.
    PATTERN RECOGNITION AND IMAGE ANALYSIS, 2023, 33 (03) : 467 - 477
  • [23] Actor-Critic based Improper Reinforcement Learning
    Zaki, Mohammadi
    Mohan, Avinash
    Gopalan, Aditya
    Mannor, Shie
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [24] Curious Hierarchical Actor-Critic Reinforcement Learning
    Roeder, Frank
    Eppe, Manfred
    Nguyen, Phuong D. H.
    Wermter, Stefan
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2020, PT II, 2020, 12397 : 408 - 419
  • [25] Applying Online Expert Supervision in Deep Actor-Critic Reinforcement Learning
    Zhang, Jin
    Chen, Jiansheng
    Huang, Yiqing
    Wan, Weitao
    Li, Tianpeng
    PATTERN RECOGNITION AND COMPUTER VISION, PT II, 2018, 11257 : 469 - 478
  • [26] Coverage Path Planning Using Actor-Critic Deep Reinforcement Learning
    Garrido-Castaneda, Sergio Isahi
    Vasquez, Juan Irving
    Antonio-Cruz, Mayra
    SENSORS, 2025, 25 (05)
  • [27] Fully distributed actor-critic architecture for multitask deep reinforcement learning
    Valcarcel Macua, Sergio
    Davies, Ian
    Tukiainen, Aleksi
    De Cote, Enrique Munoz
    KNOWLEDGE ENGINEERING REVIEW, 2021, 36
  • [28] A fuzzy Actor-Critic reinforcement learning network
    Wang, Xue-Song
    Cheng, Yu-Hu
    Yi, Jian-Qiang
    INFORMATION SCIENCES, 2007, 177 (18) : 3764 - 3781
  • [29] A modified actor-critic reinforcement learning algorithm
    Mustapha, SM
    Lachiver, G
    2000 CANADIAN CONFERENCE ON ELECTRICAL AND COMPUTER ENGINEERING, CONFERENCE PROCEEDINGS, VOLS 1 AND 2: NAVIGATING TO A NEW ERA, 2000, : 605 - 609
  • [30] Research on actor-critic reinforcement learning in RoboCup
    Guo, He
    Liu, Tianying
    Wang, Yuxin
    Chen, Feng
    Fan, Jianming
    WCICA 2006: SIXTH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION, VOLS 1-12, CONFERENCE PROCEEDINGS, 2006, : 205 - 205