Cooperative Multi-Agent Reinforcement Learning for Large Scale Variable Speed Limit Control

被引:3
|
作者
Zhang, Yuhang [1 ]
Quinones-Grueiro, Marcos [1 ]
Barbour, William [1 ]
Zhang, Zhiyao [1 ]
Scherer, Joshua [1 ]
Biswas, Gautam [1 ]
Work, Daniel [1 ]
机构
[1] Vanderbilt Univ, Inst Software Integrated Syst, Nashville, TN 37212 USA
关键词
traffic control; variable speed limit; multi-agent reinforcement learning; CONGESTION;
D O I
10.1109/SMARTCOMP58114.2023.00036
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Variable speed limit (VSL) control has emerged as a promising traffic management strategy for enhancing safety and mobility. In this study, we introduce a multi-agent reinforcement learning framework for implementing a large-scale VSL system to address recurring congestion in transportation corridors. The VSL control problem is modeled as a Markov game, using only data widely available on freeways. By employing parameter sharing among all VSL agents, the proposed algorithm can efficiently scale to cover extensive corridors. The agents are trained using a reward structure that incorporates adaptability, safety, mobility, and penalty terms; enabling agents to learn a coordinated policy that effectively reduces spatial speed variations while minimizing the impact on mobility. Our findings reveal that the proposed algorithm leads to a significant reduction in speed variation, which holds the potential to reduce incidents. Furthermore, the proposed approach performs satisfactorily under varying traffic demand and compliance rates.
引用
收藏
页码:149 / 156
页数:8
相关论文
共 50 条
  • [21] Adaptive Traffic Signal Control for large-scale scenario with Cooperative Group-based Multi-agent reinforcement learning
    Wang, Tong
    Cao, Jiahua
    Hussain, Azhar
    TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2021, 125
  • [22] Deep Reinforcement Learning Agent for Negotiation in Multi-Agent Cooperative Distributed Predictive Control
    Aponte-Rengifo, Oscar
    Vega, Pastora
    Francisco, Mario
    APPLIED SCIENCES-BASEL, 2023, 13 (04):
  • [23] Engineering A Large-Scale Traffic Signal Control: A Multi-Agent Reinforcement Learning Approach
    Chen, Yue
    Li, Changle
    Yue, Wenwei
    Zhang, Hehe
    Mao, Guoqiang
    IEEE CONFERENCE ON COMPUTER COMMUNICATIONS WORKSHOPS (IEEE INFOCOM WKSHPS 2021), 2021,
  • [24] GPLight: Grouped Multi-agent Reinforcement Learning for Large-scale Traffic Signal Control
    Liu, Yilin
    Luo, Guiyang
    Yuan, Quan
    Li, Jinglin
    Jin, Lei
    Chen, Bo
    Pan, Rui
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 199 - 207
  • [25] Learning Cooperative Intrinsic Motivation in Multi-Agent Reinforcement Learning
    Hong, Seung-Jin
    Lee, Sang-Kwang
    12TH INTERNATIONAL CONFERENCE ON ICT CONVERGENCE (ICTC 2021): BEYOND THE PANDEMIC ERA WITH ICT CONVERGENCE INNOVATION, 2021, : 1697 - 1699
  • [26] Cooperative Learning of Multi-Agent Systems Via Reinforcement Learning
    Wang, Xin
    Zhao, Chen
    Huang, Tingwen
    Chakrabarti, Prasun
    Kurths, Juergen
    IEEE TRANSACTIONS ON SIGNAL AND INFORMATION PROCESSING OVER NETWORKS, 2023, 9 : 13 - 23
  • [27] Multi-agent cooperative learning research based on reinforcement learning
    Liu, Fei
    Zeng, Guangzhou
    2006 10TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, PROCEEDINGS, VOLS 1 AND 2, 2006, : 1408 - 1413
  • [28] Cooperative Multi-Agent Reinforcement Learning With Approximate Model Learning
    Park, Young Joon
    Lee, Young Jae
    Kim, Seoung Bum
    IEEE ACCESS, 2020, 8 : 125389 - 125400
  • [29] Cooperative Multi-Agent Reinforcement Learning with Hypergraph Convolution
    Bai, Yunpeng
    Gong, Chen
    Zhang, Bin
    Fan, Guoliang
    Hou, Xinwen
    Lu, Yu
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [30] Multi-agent Cooperative Search based on Reinforcement Learning
    Sun, Yinjiang
    Zhang, Rui
    Liang, Wenbao
    Xu, Cheng
    PROCEEDINGS OF 2020 3RD INTERNATIONAL CONFERENCE ON UNMANNED SYSTEMS (ICUS), 2020, : 891 - 896