Actor-Critic Deep Reinforcement Learning for Solving Job Shop Scheduling Problems

被引:142
|
作者
Liu, Chien-Liang [1 ]
Chang, Chuan-Chin [1 ]
Tseng, Chun-Jan [1 ]
机构
[1] Natl Chiao Tung Univ, Dept Ind Engn & Management, Hsinchu 30010, Taiwan
来源
IEEE ACCESS | 2020年 / 8卷
关键词
Job shop scheduling; Machine learning; Benchmark testing; Dynamic scheduling; Learning (artificial intelligence); Training; Optimization; Job shop scheduling problem ([!text type='JS']JS[!/text]SP); deep reinforcement learning; actor-critic network; parallel training; OPTIMIZATION; SEARCH; LEVEL; GAME; GO;
D O I
10.1109/ACCESS.2020.2987820
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In the past decades, many optimization methods have been devised and applied to job shop scheduling problem (JSSP) to find the optimal solution. Many methods assumed that the scheduling results were applied to static environments, but the whole environments in the real world are always dynamic. Moreover, many unexpected events such as machine breakdowns and material problems may be present to adversely affect the initial job scheduling. This work views JSSP as a sequential decision making problem and proposes to use deep reinforcement learning to cope with this problem. The combination of deep learning and reinforcement learning avoids handcraft features as used in traditional reinforcement learning, and it is expected that the combination will make the whole learning phase more efficient. Our proposed model comprises actor network and critic network, both including convolution layers and fully connected layer. Actor network agent learns how to behave in different situations, while critic network helps agent evaluate the value of statement then return to actor network. This work proposes a parallel training method, combining asynchronous update as well as deep deterministic policy gradient (DDPG), to train the model. The whole network is trained with parallel training on a multi-agent environment and different simple dispatching rules are considered as actions. We evaluate our proposed model on more than ten instances that are present in a famous benchmark problem library - OR library. The evaluation results indicate that our method is comparative in static JSSP benchmark problems, and achieves a good balance between makespan and execution time in dynamic environments. Scheduling score of our method is 91.12% in static JSSP benchmark problems, and 80.78% in dynamic environments.
引用
收藏
页码:71752 / 71762
页数:11
相关论文
共 50 条
  • [41] Variational value learning in advantage actor-critic reinforcement learning
    Zhang, Yaozhong
    Han, Jiaqi
    Hu, Xiaofang
    Dan, Shihao
    2020 CHINESE AUTOMATION CONGRESS (CAC 2020), 2020, : 1955 - 1960
  • [42] Estimation Error Correction in Deep Reinforcement Learning for Deterministic Actor-Critic Methods
    Saglam, Baturay
    Duran, Enes
    Cicek, Dogan C.
    Mutlu, Furkan B.
    Kozat, Suleyman S.
    2021 IEEE 33RD INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2021), 2021, : 137 - 144
  • [43] Dynamic spectrum access and sharing through actor-critic deep reinforcement learning
    Dong, Liang
    Qian, Yuchen
    Xing, Yuan
    EURASIP JOURNAL ON WIRELESS COMMUNICATIONS AND NETWORKING, 2022, 2022 (01)
  • [44] Solutions to Finite Horizon Cost Problems Using Actor-Critic Reinforcement Learning
    Grondman, Ivo
    Xu, Hao
    Jagannathan, Sarangapani
    Babuska, Robert
    2013 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2013,
  • [45] Symmetric actor-critic deep reinforcement learning for cascade quadrotor flight control
    Han, Haoran
    Cheng, Jian
    Xi, Zhilong
    Lv, Maolong
    NEUROCOMPUTING, 2023, 559
  • [46] Stochastic Latent Actor-Critic: Deep Reinforcement Learning with a Latent Variable Model
    Lee, Alex X.
    Nagabandi, Anusha
    Abbeel, Pieter
    Levine, Sergey
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [47] Actor-Critic Reinforcement Learning for Tracking Control in Robotics
    Pane, Yudha P.
    Nageshrao, Subramanya P.
    Babuska, Robert
    2016 IEEE 55TH CONFERENCE ON DECISION AND CONTROL (CDC), 2016, : 5819 - 5826
  • [48] Reinforcement learning with actor-critic for knowledge graph reasoning
    Zhang, Linli
    Li, Dewei
    Xi, Yugeng
    Jia, Shuai
    SCIENCE CHINA-INFORMATION SCIENCES, 2020, 63 (06)
  • [49] Actor-critic reinforcement learning for bidding in bilateral negotiation
    Arslan, Furkan
    Aydogan, Reyhan
    TURKISH JOURNAL OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCES, 2022, 30 (05) : 1695 - 1714
  • [50] Reinforcement learning with actor-critic for knowledge graph reasoning
    Linli Zhang
    Dewei Li
    Yugeng Xi
    Shuai Jia
    Science China Information Sciences, 2020, 63