XLight: An interpretable multi-agent reinforcement learning approach for traffic signal control

被引:0
|
作者
Cai, Sibin [1 ]
Fang, Jie [1 ]
Xu, Mengyun [1 ,2 ]
机构
[1] Fuzhou Univ, Coll Civil Engn, Fuzhou 350108, Peoples R China
[2] Natl Univ Singapore, Dept Civil & Environm Engn, Singapore 119077, Singapore
基金
中国国家自然科学基金;
关键词
Multi-agent reinforcement learning; Traffic signal control; Interpretability; Regulatable function; Maximum entropy policy optimization;
D O I
10.1016/j.eswa.2025.126938
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, deep reinforcement learning (DRL)-based traffic signal control (TSC) methods have garnered significant attention among researchers, achieving substantial progress. However, current research often focuses on performance improvement, neglecting interpretability. DRL-based TSC methods often face challenges in interpretability. This limitation poses significant obstacles to practical deployment, given the liability and regulatory constraints faced by governmental authorities responsible for traffic management and control. On the other hand, interpretable RL-based TSC methods offer greater flexibility to meet specific requirements. For instance, prioritizing the clearance of vehicles in a particular movement can be easily achieved by assigning higher weights to the state variables associated with that movement. To address this issue, we propose Xlight, an interpretable multi-agent reinforcement learning (MARL) approach for TSC, which enhances interpretability in three key aspects: (a) meticulously designing and selecting the state space, action space, and reward function. Especially, we propose an interpretable reward function for network-wide TSC and prove that maximizing this reward is equivalent to minimizing the average travel time (ATT) in the road network; (b) introducing more practical regulatable (i.e., interpretable) functions as TSC controllers; and (c) employing maximum entropy policy optimization, which simultaneously enhances interpretability and improves transferability. Next, to better align with practical applications of network-wide TSC, we propose several interpretable MARL-based methods. Among these, Multi-Agent Regulatable Soft Actor-Critic (MARSAC) not only possesses interpretability but also achieves superior performance. Finally, comprehensive experiments conducted across various TSC scenarios, including isolated intersection, synthetic network-wide intersections, and real-world network-wide intersections, demonstrate the effectiveness. For example, in terms of the ATT metric, our proposed method achieves improvements of 9.55%, 34.17%, 3.98%, and 42.93% compared to the Actuated Traffic Signal Control (ATSC) across a synthetic road network and 3 real-world road networks. Furthermore, in the synthetic network, our method demonstrates improvements of 4.04% and 3.21% in the Safety Score and Fuel Consumption metrics, respectively, when compared to the ATSC.
引用
收藏
页数:20
相关论文
共 50 条
  • [31] GPLight: Grouped Multi-agent Reinforcement Learning for Large-scale Traffic Signal Control
    Liu, Yilin
    Luo, Guiyang
    Yuan, Quan
    Li, Jinglin
    Jin, Lei
    Chen, Bo
    Pan, Rui
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 199 - 207
  • [32] AGRCNet: communicate by attentional graph relations in multi-agent reinforcement learning for traffic signal control
    Tinghuai Ma
    Kexing Peng
    Huan Rong
    Yurong Qian
    Neural Computing and Applications, 2023, 35 : 21007 - 21022
  • [33] Traffic signal priority control based on shared experience multi-agent deep reinforcement learning
    Wang, Zhiwen
    Yang, Kangkang
    Li, Long
    Lu, Yanrong
    Tao, Yufei
    IET INTELLIGENT TRANSPORT SYSTEMS, 2023, 17 (07) : 1363 - 1379
  • [34] Multi-Agent Meta-Reinforcement Learning with Coordination and Reward Shaping for Traffic Signal Control
    Du, Xin
    Wang, Jiahai
    Chen, Siyuan
    ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2023, PT II, 2023, 13936 : 349 - 360
  • [35] Traffic signal control using a cooperative EWMA-based multi-agent reinforcement learning
    Qiao, Zhimin
    Ke, Liangjun
    Wang, Xiaoqiang
    APPLIED INTELLIGENCE, 2023, 53 (04) : 4483 - 4498
  • [36] A Distributed Multi-Agent Reinforcement Learning With Graph Decomposition Approach for Large-Scale Adaptive Traffic Signal Control
    Jiang, Shan
    Huang, Yufei
    Jafari, Mohsen
    Jalayer, Mohammad
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (09) : 14689 - 14701
  • [37] Traffic Optimization in Satellites Communications: A Multi-agent Reinforcement Learning Approach
    Qin, Zeyu
    Yao, Haipeng
    Mai, Tianle
    2020 16TH INTERNATIONAL WIRELESS COMMUNICATIONS & MOBILE COMPUTING CONFERENCE, IWCMC, 2020, : 269 - 273
  • [38] Multi-agent broad reinforcement learning for intelligent traffic light control
    Zhu, Ruijie
    Li, Lulu
    Wu, Shuning
    Lv, Pei
    Li, Yafei
    Xu, Mingliang
    INFORMATION SCIENCES, 2023, 619 : 509 - 525
  • [39] Regional Multi-Agent Cooperative Reinforcement Learning for City-Level Traffic Grid Signal Control
    Yisha Li
    Ya Zhang
    Xinde Li
    Changyin Sun
    IEEE/CAAJournalofAutomaticaSinica, 2024, 11 (09) : 1987 - 1998
  • [40] Multi-agent Deep Reinforcement Learning collaborative Traffic Signal Control method considering intersection heterogeneity
    Bie, Yiming
    Ji, Yuting
    Ma, Dongfang
    TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2024, 164