Shunting Trains with Deep Reinforcement Learning

被引:18
|
作者
Peer, Evertjan [1 ]
Menkovski, Vlado [1 ]
Zhang, Yingqian [1 ]
Lee, Wan-Jui [2 ]
机构
[1] Eindhoven Univ Technol, Eindhoven, Netherlands
[2] NS Dutch Railways, Maintenance Dev, Utrecht, Netherlands
基金
欧盟地平线“2020”;
关键词
D O I
10.1109/SMC.2018.00520
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
The Train Unit Shunting Problem (TUSP) is a difficult sequential decision making problem faced by Dutch Railways (NS). Current heuristic solutions under study at NS fall short in accounting for uncertainty during plan execution and do not efficiently support replanning. Furthermore, the resulting plans lack consistency. We approach the TUSP by formulating it as a Markov Decision Process and develop an image-like state space representation that allows us to develop a Deep Reinforcement Learning (DRL) solution. The Deep Q-Network efficiently reduces the state space and develops an on-line strategy for the TUSP capable of dealing with uncertainty and delivering significantly more consistent solutions compared to approaches currently being developed by NS.
引用
收藏
页码:3063 / 3068
页数:6
相关论文
共 50 条
  • [41] Learning Global Optimization by Deep Reinforcement Learning
    da Silva Filho, Moesio Wenceslau
    Barbosa, Gabriel A.
    Miranda, Pericles B. C.
    INTELLIGENT SYSTEMS, PT II, 2022, 13654 : 417 - 433
  • [42] Learning Macromanagement in Starcraft by Deep Reinforcement Learning
    Huang, Wenzhen
    Yin, Qiyue
    Zhang, Junge
    Huang, Kaiqi
    SENSORS, 2021, 21 (10)
  • [43] Deep sparse representation via deep dictionary learning for reinforcement learning
    Tang, Jianhao
    Li, Zhenni
    Xie, Shengli
    Ding, Shuxue
    Zheng, Shaolong
    Chen, Xueni
    2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 2398 - 2403
  • [44] Optimization of Shunting Operation Plan for Detaching and Attaching Trains Based on Q-Learning Algorithm
    Shi J.
    Chen L.
    Lin B.
    Meng G.
    Xia S.
    Zhongguo Tiedao Kexue/China Railway Science, 2022, 43 (01): : 163 - 170
  • [45] Towards Interpretable Deep Reinforcement Learning Models via Inverse Reinforcement Learning
    Xie, Yuansheng
    Vosoughi, Soroush
    Hassanpour, Saeed
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 5067 - 5074
  • [46] Asynchronous Methods for Deep Reinforcement Learning
    Mnih, Volodymyr
    Badia, Adria Puigdomenech
    Mirza, Mehdi
    Graves, Alex
    Harley, Tim
    Lillicrap, Timothy P.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48
  • [47] Neuroevolution for Deep Reinforcement Learning Problems
    Ha, David
    PROCEEDINGS OF THE 2019 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE COMPANION (GECCCO'19 COMPANION), 2019, : 550 - 593
  • [48] Shallow Updates for Deep Reinforcement Learning
    Levine, Nir
    Zahavy, Tom
    Mankowitz, Daniel J.
    Tamar, Aviv
    Mannor, Shie
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [49] Deep reinforcement learning of transition states
    Zhang, Jun
    Lei, Yao-Kun
    Zhang, Zhen
    Han, Xu
    Li, Maodong
    Yang, Lijiang
    Yang, Yi Isaac
    Gao, Yi Qin
    PHYSICAL CHEMISTRY CHEMICAL PHYSICS, 2021, 23 (11) : 6888 - 6895
  • [50] Deep Reinforcement Learning with Hierarchical Structures
    Li, Siyuan
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 4899 - 4900