Shunting Trains with Deep Reinforcement Learning

被引:18
|
作者
Peer, Evertjan [1 ]
Menkovski, Vlado [1 ]
Zhang, Yingqian [1 ]
Lee, Wan-Jui [2 ]
机构
[1] Eindhoven Univ Technol, Eindhoven, Netherlands
[2] NS Dutch Railways, Maintenance Dev, Utrecht, Netherlands
基金
欧盟地平线“2020”;
关键词
D O I
10.1109/SMC.2018.00520
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
The Train Unit Shunting Problem (TUSP) is a difficult sequential decision making problem faced by Dutch Railways (NS). Current heuristic solutions under study at NS fall short in accounting for uncertainty during plan execution and do not efficiently support replanning. Furthermore, the resulting plans lack consistency. We approach the TUSP by formulating it as a Markov Decision Process and develop an image-like state space representation that allows us to develop a Deep Reinforcement Learning (DRL) solution. The Deep Q-Network efficiently reduces the state space and develops an on-line strategy for the TUSP capable of dealing with uncertainty and delivering significantly more consistent solutions compared to approaches currently being developed by NS.
引用
收藏
页码:3063 / 3068
页数:6
相关论文
共 50 条
  • [31] Deep Reinforcement Learning for Adaptive Learning Systems
    Li, Xiao
    Xu, Hanchen
    Zhang, Jinming
    Chang, Hua-hua
    JOURNAL OF EDUCATIONAL AND BEHAVIORAL STATISTICS, 2023, 48 (02) : 220 - 243
  • [32] Learning to Walk via Deep Reinforcement Learning
    Haarnoja, Tuomas
    Ha, Sehoon
    Zhou, Aurick
    Tan, Jie
    Tucker, George
    Levine, Sergey
    ROBOTICS: SCIENCE AND SYSTEMS XV, 2019,
  • [33] Learning to Break Rocks With Deep Reinforcement Learning
    Samtani, Pavan
    Leiva, Francisco
    Ruiz-del-Solar, Javier
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (02) : 1077 - 1084
  • [34] The Difficulty of Passive Learning in Deep Reinforcement Learning
    Ostrovski, Georg
    Castro, Pablo Samuel
    Dabney, Will
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [35] Deep learning and reinforcement learning approach on microgrid
    Chandrasekaran, Kumar
    Kandasamy, Prabaakaran
    Ramanathan, Srividhya
    INTERNATIONAL TRANSACTIONS ON ELECTRICAL ENERGY SYSTEMS, 2020, 30 (10):
  • [36] Contrastive Learning Methods for Deep Reinforcement Learning
    Wang, Di
    Hu, Mengqi
    IEEE ACCESS, 2023, 11 : 97107 - 97117
  • [37] Transfer Learning in Deep Reinforcement Learning: A Survey
    Zhu, Zhuangdi
    Lin, Kaixiang
    Jain, Anil K.
    Zhou, Jiayu
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (11) : 13344 - 13362
  • [38] Learning an Index Advisor with Deep Reinforcement Learning
    Lai, Sichao
    Wu, Xiaoying
    Wang, Senyang
    Peng, Yuwei
    Peng, Zhiyong
    WEB AND BIG DATA, APWEB-WAIM 2021, PT II, 2021, 12859 : 178 - 185
  • [39] Deep learning, reinforcement learning, and world models
    Matsuo, Yutaka
    LeCun, Yann
    Sahani, Maneesh
    Precup, Doina
    Silver, David
    Sugiyama, Masashi
    Uchibe, Eiji
    Morimoto, Jun
    NEURAL NETWORKS, 2022, 152 : 267 - 275
  • [40] Learning Dynamics and Generalization in Deep Reinforcement Learning
    Lyle, Clare
    Rowland, Mark
    Dabney, Will
    Kwiatkowksa, Marta
    Gal, Yarin
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,