Improving Traffic Signal Control With Joint-Action Reinforcement Learning

被引:0
|
作者
Labres, Joao V. B. [1 ]
Bazzan, Ana L. C. [1 ]
Abdoos, Monireh [2 ]
机构
[1] Univ Fed Rio Grande do Sul, Inst Informat, Porto Alegre, RS, Brazil
[2] Shahid Beheshti Univ, Fac Comp Sci & Engn, Tehran, Iran
关键词
reinforcement learning; traffic signal control;
D O I
10.1109/SSCI50451.2021.9659871
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
An increasing demand for mobility in urban areas leads to several issues such as traffic congestion. While many approaches exist to tackle this problem, a popular one is the adaptive control of traffic signals. This paper focuses on reinforcement learning based approaches for adaptive control. These use learning agents that act individually to adapt the signal timings to the current demand of vehicles crossing an intersection. However, in a network of intersections (thus constituting a multiagent reinforcement learning problem), as the actions of each agent have impact on the states, actions, and rewards of others, it is clear that there is a relevant non-local issue that is rarely dealt with. Thus, this paper proposes a joint learning approach in which agents learn in groups, besides also learning an individual policy. This means that we have two parallel learning processes. The group policy is implemented from time to time, thus helping agents to implicitly consider other agents' policies. A distinguishing feature of our approach is that it is not based on fixed membership in such groups; rather, groups are formed and dissolved on demand, based on their performance. Results of applying this approach are compared to a baseline and to when agents learn individually only, showing that the joint learning leads to lesser waiting time.
引用
收藏
页数:8
相关论文
共 50 条
  • [1] Leveraging Joint-Action Embedding in Multiagent Reinforcement Learning for Cooperative Games
    Lou, Xingzhou
    Zhang, Junge
    Du, Yali
    Yu, Chao
    He, Zhaofeng
    Huang, Kaiqi
    IEEE TRANSACTIONS ON GAMES, 2024, 16 (02) : 470 - 482
  • [2] Reinforcement Learning for Traffic Signal Control in Hybrid Action Space
    Luo, Haoqing
    Bie, Yiming
    Jin, Sheng
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, 25 (06) : 5225 - 5241
  • [3] Mitigating Action Hysteresis in Traffic Signal Control with Traffic Predictive Reinforcement Learning
    Han, Xiao
    Zhao, Xiangyu
    Zhang, Liang
    Wang, Wanyu
    PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023, 2023, : 673 - 684
  • [4] LJIR: Learning Joint-Action Intrinsic Reward in cooperative multi-agent reinforcement learning
    Chen, Zihan
    Luo, Biao
    Hu, Tianmeng
    Xu, Xiaodong
    NEURAL NETWORKS, 2023, 167 : 450 - 459
  • [5] Baselines for joint-action reinforcement learning of coordination in cooperative multi-agent systems
    Carpenter, M
    Kudenko, D
    ADAPTIVE AGENTS AND MULTI-AGENT SYSTEMS II: ADAPTATION AND MULTI-AGENT LEARNING, 2005, 3394 : 55 - 72
  • [6] Traffic Signal Control Using Hybrid Action Space Deep Reinforcement Learning
    Bouktif, Salah
    Cheniki, Abderraouf
    Ouni, Ali
    SENSORS, 2021, 21 (07)
  • [7] Parallel Reinforcement Learning for Traffic Signal Control
    Mannion, Patrick
    Duggan, Jim
    Howley, Enda
    6TH INTERNATIONAL CONFERENCE ON AMBIENT SYSTEMS, NETWORKS AND TECHNOLOGIES (ANT-2015), THE 5TH INTERNATIONAL CONFERENCE ON SUSTAINABLE ENERGY INFORMATION TECHNOLOGY (SEIT-2015), 2015, 52 : 956 - 961
  • [8] Reinforcement Learning with Explainability for Traffic Signal Control
    Rizzo, Stefano Giovanni
    Vantini, Giovanna
    Chawla, Sanjay
    2019 IEEE INTELLIGENT TRANSPORTATION SYSTEMS CONFERENCE (ITSC), 2019, : 3567 - 3572
  • [9] Traffic Signal Control Using Reinforcement Learning
    Jadhao, Namrata S.
    Jadhao, Ashish S.
    2014 FOURTH INTERNATIONAL CONFERENCE ON COMMUNICATION SYSTEMS AND NETWORK TECHNOLOGIES (CSNT), 2014, : 1130 - 1135
  • [10] Reinforcement learning in neurofuzzy traffic signal control
    Bingham, E
    EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 2001, 131 (02) : 232 - 241