EGLight: enhancing deep reinforcement learning with expert guidance for traffic signal control

被引:0
|
作者
Zhang, Meng [1 ,2 ]
Wang, Dianhai [1 ,3 ]
Cai, Zhengyi [4 ]
Huang, Yulang [1 ]
Yu, Hongxin [1 ]
Qin, Hanwu [1 ]
Zeng, Jiaqi [1 ,3 ]
机构
[1] Zhejiang Univ, Coll Civil Engn & Architecture, Hangzhou 310058, Peoples R China
[2] Zhejiang Urban Governance Studies Ctr, Hangzhou, Peoples R China
[3] Zhejiang Univ, Zhongyuan Inst, Zhengzhou 450001, Peoples R China
[4] Hangzhou City Univ, Sch Informat & Elect Engn, Hangzhou, Peoples R China
基金
中国国家自然科学基金;
关键词
Traffic signal control; deep reinforcement learning; supervised learning; expert policy; LIGHT CONTROL; GRADIENT;
D O I
10.1080/23249935.2025.2486263
中图分类号
U [交通运输];
学科分类号
08 ; 0823 ;
摘要
Deep Reinforcement Learning (DRL) is prevalent in traffic signal control. However, the training process often encounters slow learning rate and unstable convergence due to limited state representation and exploratory learning. Inspired by human learning, we incorporate expert guidance in the exploration process to accelerate convergence and enhance performance. The proposed framework, termed Expert-Guided Light (EGLight), contains three moudles. The state perception module combines statistical features with cellular features to enhance model robustness. The decision-making module employs expert-guided learning to promote the learning efficiency. In the learning module, four distinct loss functions are employed to make full use of the interaction experience and update the agent's policy. Extensive tests demonstrate EGLight's superior convergence speed and effectiveness over traditional methods. The analysis shows that the precise feature design is helpful for the agent and proper expert guidance is crucial for the convergence of agent learning process.
引用
收藏
页数:27
相关论文
共 50 条
  • [31] Minimize Pressure Difference Traffic Signal Control Based on Deep Reinforcement Learning
    Yu, Pengcheng
    Luo, Jie
    2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 5493 - 5498
  • [32] Adaptive urban traffic signal control based on enhanced deep reinforcement learning
    Cai, Changjian
    Wei, Min
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [33] Adaptive Traffic Signal Control Model on Intersections Based on Deep Reinforcement Learning
    Li, Duowei
    Wu, Jianping
    Xu, Ming
    Wang, Ziheng
    Hu, Kezhen
    Journal of Advanced Transportation, 2020, 2020
  • [34] Adaptive Traffic Signal Control Model on Intersections Based on Deep Reinforcement Learning
    Li, Duowei
    Wu, Jianping
    Xu, Ming
    Wang, Ziheng
    Hu, Kezhen
    JOURNAL OF ADVANCED TRANSPORTATION, 2020, 2020
  • [35] Traffic Signal Control with Deep Reinforcement Learning and Self-attention Mechanism
    Zhang X.
    Nie S.
    Li Z.
    Zhang H.
    Jiaotong Yunshu Xitong Gongcheng Yu Xinxi/Journal of Transportation Systems Engineering and Information Technology, 2024, 24 (02): : 96 - 104
  • [36] Application of Deep Reinforcement Learning in Traffic Signal Control: An Overview and Impact of Open Traffic Data
    Greguric, Martin
    Vujic, Miroslav
    Alexopoulos, Charalampos
    Miletic, Mladen
    APPLIED SCIENCES-BASEL, 2020, 10 (11):
  • [37] Parallel Reinforcement Learning for Traffic Signal Control
    Mannion, Patrick
    Duggan, Jim
    Howley, Enda
    6TH INTERNATIONAL CONFERENCE ON AMBIENT SYSTEMS, NETWORKS AND TECHNOLOGIES (ANT-2015), THE 5TH INTERNATIONAL CONFERENCE ON SUSTAINABLE ENERGY INFORMATION TECHNOLOGY (SEIT-2015), 2015, 52 : 956 - 961
  • [38] Reinforcement Learning with Explainability for Traffic Signal Control
    Rizzo, Stefano Giovanni
    Vantini, Giovanna
    Chawla, Sanjay
    2019 IEEE INTELLIGENT TRANSPORTATION SYSTEMS CONFERENCE (ITSC), 2019, : 3567 - 3572
  • [39] Reinforcement learning in neurofuzzy traffic signal control
    Bingham, E
    EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 2001, 131 (02) : 232 - 241
  • [40] Traffic Signal Control Using Reinforcement Learning
    Jadhao, Namrata S.
    Jadhao, Ashish S.
    2014 FOURTH INTERNATIONAL CONFERENCE ON COMMUNICATION SYSTEMS AND NETWORK TECHNOLOGIES (CSNT), 2014, : 1130 - 1135