Multi-Agent Deep Reinforcement Learning with Clustering and Information Sharing for Traffic Light Cooperative Control

被引:0
|
作者
Du T. [1 ]
Wang B. [1 ]
Cheng H. [1 ]
Luo L. [1 ]
Zeng N. [2 ]
机构
[1] School of Computer and Information, Anhui Normal University, Wuhu
[2] College of Economics and Management, Harbin Engineering University, Harbin
关键词
Centralized training with decentralized execution; Deep recurrent Q-network; Growing neural gas; Reinforcement learning agent cluster; Traffic light cooperative control;
D O I
10.11999/JEIT230857
中图分类号
学科分类号
摘要
In order to improve the joint control effect of multi-crossing, Multi-Agent Deep Recurrent Q-Network (MADRQN) for real-time control of multi-intersection traffic signals is proposed in this paper. Firstly, the traffic light control is modeled as a Markov decision process, wherein one controller at each crossing is considered as an agent. Secondly, agents are clustered according to their position and observation. Then, information sharing and centralized training are conducted within each cluster. Also the value function network parameters of agents with the highest critic value are shared with other agent at the end of every training process. The simulated experimental results under Simulation of Urban MObility (SUMO) show that the proposed method can reduce the amount of communication data, make information sharing of agents and centralized training more feasible and efficient. The average delay of vehicles is reduced obviously compared with the state-of-the-art traffic light control methods based on multi-agent deep reinforcement learning. The proposed method can effectively alleviate traffic congestion. © 2024 Science Press. All rights reserved.
引用
收藏
页码:538 / 545
页数:7
相关论文
共 21 条
  • [1] PANDIT K, GHOSAL D, ZHANG H M, Et al., Adaptive traffic signal control with vehicular ad hoc networks[J], IEEE Transactions on Vehicular Technology, 62, 4, pp. 1459-1471, (2013)
  • [2] HE Kaiming, ZHANG Xiangyu, REN Shaoqing, Et al., Deep residual learning for image recognition[C], 2016 IEEE Conference on Computer Vision and Pattern Recognition, pp. 770-778, (2016)
  • [3] SHAO Mingli, CAO E, HU Ming, Et al., Traffic light optimization control method for priority vehicle awareness[J], Journal of Software, 32, 8, pp. 2425-2438, (2021)
  • [4] HADDAD T A, HEDJAZI D, AOUAG S., A new deep reinforcement learning-based adaptive traffic light control approach for isolated intersection[C], The 5th International Symposium on Informatics and its Applications, pp. 1-6, (2022)
  • [5] GENDERS W, RAZAVI S., Using a deep reinforcement learning agent for traffic signal control, (2016)
  • [6] TIGGA A, HOTA L, PATEL S, Et al., A deep Q-learning-based adaptive traffic light control system for urban safety[C], The 4th International Conference on Advances in Computing, Communication Control and Networking, pp. 2430-2435, (2022)
  • [7] ZOU Xiangyu, HUANG Chongwen, XU Yongjun, Et al., Secure energy efficiency in communication systems based on deep learning[J], Journal of Electronics & Information Technology, 44, 7, pp. 2245-2252, (2022)
  • [8] TANG Lun, LI Zhixuan, PU Hao, Et al., A dynamic predeployment strategy of UAVs based on multi-agent deep reinforcement learning[J], Journal of Electronics & Information Technology, 45, 6, pp. 2007-2015, (2023)
  • [9] KANG Leilei, HUANG Hao, LU Weike, Et al., A dueling deep Q-network method for low-carbon traffic signal control, Applied Soft Computing, 141, (2023)
  • [10] TUNC I, SOYLEMEZ M T., Fuzzy logic and deep Q learning based control for traffic lights[J], Alexandria Engineering Journal, 67, pp. 343-359, (2023)