Cooperative bus holding and stop-skipping: A deep reinforcement learning framework

被引:6
|
作者
Rodriguez, Joseph [1 ,3 ]
Koutsopoulos, Haris N. [1 ]
Wang, Shenhao [2 ]
Zhao, Jinhua [2 ]
机构
[1] Northeastern Univ, 360 Huntington Ave, Boston, MA 02115 USA
[2] MIT, 77 Massachusetts Ave, Cambridge, MA 02139 USA
[3] Snell Engn, 110 Forsyth St, Boston, MA 02115 USA
关键词
Real-time bus control; Multi-agent reinforcement learning; Cooperative agents; TRANSIT; MODEL; SIMULATION;
D O I
10.1016/j.trc.2023.104308
中图分类号
U [交通运输];
学科分类号
08 ; 0823 ;
摘要
The bus control problem that combines holding and stop-skipping strategies is formulated as a multi-agent reinforcement learning (MARL) problem. Traditional MARL methods, designed for settings with joint action-taking, are incompatible with the asynchronous nature of at-stop control tasks. On the other hand, using a fully decentralized approach leads to environment non-stationarity, since the state transition of an individual agent may be distorted by the actions of other agents. To address it, we propose a design of the state and reward function that increases the observability of the impact of agents' actions during training. An event-based mesoscopic simulation model is built to train the agents. We evaluate the proposed approach in a case study with a complex route from the Chicago transit network. The proposed method is compared to a standard headway-based control and a policy trained with MARL but with no cooperative learning. The results show that the proposed method not only improves level of service but it is also more robust towards uncertainties in operations such as travel times and operator compliance with the recommended action.
引用
收藏
页数:20
相关论文
共 50 条
  • [21] A Multi-Agent Reinforcement Learning approach for bus holding control strategies
    Chen, C.X.
    Chen, W.Y.
    Chen, Z.Y.
    Advances in Transportation Studies, 2015, 2 : 41 - 54
  • [22] Trajectory optimization algorithm of skipping missile based on deep reinforcement learning
    Gong K.
    Wei H.
    Li J.
    Song X.
    Li Y.
    Li Y.
    Zhang Y.
    Beijing Hangkong Hangtian Daxue Xuebao/Journal of Beijing University of Aeronautics and Astronautics, 2023, 49 (06): : 1383 - 1393
  • [23] Learning Cooperative Visual Dialog Agents with Deep Reinforcement Learning
    Das, Abhishek
    Kottur, Satwik
    Moura, Jose M. F.
    Lee, Stefan
    Batra, Dhruv
    2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 2970 - 2979
  • [24] Deep Reinforcement Learning based dynamic optimization of bus timetable
    Ai, Guanqun
    Zuo, Xingquan
    Chen, Gang
    Wu, Binglin
    APPLIED SOFT COMPUTING, 2022, 131
  • [25] Cooperative Perception with Deep Reinforcement Learning for Connected Vehicles
    Aoki, Shunsuke
    Higuchi, Takamasa
    Altintas, Onur
    2020 IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2020, : 328 - 334
  • [26] Decentralized Scheduling for Cooperative Localization With Deep Reinforcement Learning
    Peng, Bile
    Seco-Granados, Gonzalo
    Steinmetz, Erik
    Frohle, Markus
    Wymeersch, Henk
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2019, 68 (05) : 4295 - 4305
  • [27] Cooperative Emergent Swarming Through Deep Reinforcement Learning
    Lin, Tony X.
    Lofaro, Daniel M.
    Sofge, Donald A.
    2020 IEEE 16TH INTERNATIONAL CONFERENCE ON CONTROL & AUTOMATION (ICCA), 2020, : 1354 - 1359
  • [28] Cooperative Proactive Eavesdropping Based on Deep Reinforcement Learning
    Yang, Yaxin
    Li, Baogang
    Zhang, Shue
    Zhao, Wei
    Zhang, Haijun
    IEEE WIRELESS COMMUNICATIONS LETTERS, 2021, 10 (09) : 1857 - 1861
  • [29] Determining Bus Stop Locations using Deep Learning and Time Filtering
    Piriyataravet, Jitpinun
    Kumwilaisak, Wuttipong
    Chinrungrueng, Jatuporn
    Piriyatharawet, Teerawat
    ENGINEERING JOURNAL-THAILAND, 2021, 25 (08): : 163 - 172
  • [30] A Deep Reinforcement Learning Framework for Column Generation
    Chi, Cheng
    Aboussalah, Amine Mohamed
    Khalil, Elias B.
    Wang, Juyoung
    Sherkat-Masoumi, Zoha
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,