Traffic pattern-aware elevator dispatching via deep reinforcement learning

被引:1
|
作者
Wan, Jiansong [1 ]
Lee, Kanghoon [1 ]
Shin, Hayong [1 ]
机构
[1] Korea Adv Inst Sci & Technol, Dept Ind & Syst Engn, 291 Daehak Ro, Daejeon 34141, South Korea
基金
新加坡国家研究基金会;
关键词
Elevator dispatching; Semi-Markov decision process; Deep reinforcement learning; Traffic pattern awareness; GENETIC ALGORITHM;
D O I
10.1016/j.aei.2024.102497
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This study addresses the elevator dispatching problem using deep reinforcement learning, with a specific emphasis on traffic pattern awareness. Previous studies on reinforcement learning-based elevator dispatching have largely focused on training separate models for single traffic patterns, such as up-peak, down-peak, lunchpeak, and inter-floor. This separate training approach not only introduces practical complexities by requiring an auxiliary model to predict traffic patterns for guiding dispatching decisions but is also computationally burdensome. In contrast, our goal is to develop a unified, traffic pattern-aware dispatching model. We formulate the elevator dispatching problem as a Semi-Markov Decision Process (SMDP) with novel state representation, action space, and reward function designs. To solve the formulated SMDP, we propose a Dueling Double Deep Q-Network (D3QN) architecture associated with the training algorithm. To ensure traffic pattern awareness, we train our model in a unified 'All in One' traffic scenario, employing two practical techniques to enhance the training process: (1) temporal grouping with gradient surgery and (2) incorporation of passenger arrival information. Empirical evaluations confirm the superiority of our model over multiple benchmarks, including those relying on separate, pattern-specific models. Remarkably, our unified model demonstrates robust performance across unseen traffic scenarios and performs exceptionally well in single traffic patterns despite being trained solely on the unified 'All in One' scenario. The short inference time for decision-making further solidifies the model's practical viability. Additionally, the incremental benefits contributed by each of our introduced techniques are also investigated. Our code is available at https://github.com/jswan95/RLbased-traffic-pattern-aware-elevator-dispatching
引用
收藏
页数:13
相关论文
共 50 条
  • [31] Dispatching AGVs with battery constraints using deep reinforcement learning
    Singh, Nitish
    Akcay, Alp
    Dang, Quang-Vinh
    Martagan, Tugce
    Adan, Ivo
    COMPUTERS & INDUSTRIAL ENGINEERING, 2024, 187
  • [32] A Deep Reinforcement Learning Approach for Microgrid Energy Transmission Dispatching
    Chen, Shuai
    Liu, Jian
    Cui, Zhenwei
    Chen, Zhiyu
    Wang, Hua
    Xiao, Wendong
    APPLIED SCIENCES-BASEL, 2024, 14 (09):
  • [33] Deep Reinforcement Learning for Ride-sharing Dispatching and Repositioning
    Qin, Zhiwei
    Tang, Xiaocheng
    Jiao, Yan
    Zhang, Fan
    Wang, Chenxi
    Li, Qun
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 6566 - 6568
  • [34] Intelligent Traffic Light via Policy-based Deep Reinforcement Learning
    Zhu, Yue
    Cai, Mingyu
    Schwarz, Chris W.
    Li, Junchao
    Xiao, Shaoping
    INTERNATIONAL JOURNAL OF INTELLIGENT TRANSPORTATION SYSTEMS RESEARCH, 2022, 20 (03) : 734 - 744
  • [35] Smart dispatching and optimal elevator group control through real-time occupancy-aware deep learning of usage patterns
    Wang, Shu
    Gong, Xuejian
    Song, Mulang
    Fei, Cindy Y.
    Quaadgras, Stefan
    Peng, Jianyuan
    Zou, Pan
    Chen, Jerred
    Zhang, Wei
    Jiao, Roger J.
    ADVANCED ENGINEERING INFORMATICS, 2021, 48
  • [36] Intelligent Traffic Light via Policy-based Deep Reinforcement Learning
    Yue Zhu
    Mingyu Cai
    Chris W. Schwarz
    Junchao Li
    Shaoping Xiao
    International Journal of Intelligent Transportation Systems Research, 2022, 20 : 734 - 744
  • [37] Socially-Aware Traffic Scheduling for Edge-Assisted Metaverse by Deep Reinforcement Learning
    Yu, Ao
    Yang, Hui
    Feng, Cuiyang
    Li, Yunbo
    Zhao, Yang
    Cheriet, Mohamed
    Vasilakos, Athanasios V.
    IEEE NETWORK, 2023, 37 (06): : 74 - 81
  • [38] Deep Reinforcement Learning based Traffic- and Channel-aware OFDMA Resource Allocation
    Balakrishnan, Ravikumar
    Sankhe, Kunal
    Somayazulu, V. Srinivasa
    Vannithamby, Rath
    Sydir, Jerry
    2019 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2019,
  • [39] Patient-Specific Preictal Pattern-Aware Epileptic Seizure Prediction with Federated Learning
    Saemaldahr, Raghdah
    Ilyas, Mohammad
    SENSORS, 2023, 23 (14)
  • [40] Framework for Control and Deep Reinforcement Learning in Traffic
    Wu, Cathy
    Parvate, Kanaad
    Kheterpal, Nishant
    Dickstein, Leah
    Mehta, Ankur
    Vinitsky, Eugene
    Bayen, Alexandre M.
    2017 IEEE 20TH INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2017,