Kernel-based Reinforcement Learning for Traffic Signal Control with Adaptive Feature Selection

被引:0
|
作者
Chu, Tianshu [1 ]
Wang, Jie [1 ]
Cao, Jian [2 ]
机构
[1] Stanford Univ, Dept Civil & Environm Engn, Stanford, CA 94305 USA
[2] Shanghai Jiao Tong Univ, Dept Comp Sci & Engn, Shanghai 200030, Peoples R China
关键词
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Reinforcement learning in a large-scale system is computationally challenging due to the curse of the dimensionality. One approach is to approximate the Q-function as a function of a state-action related feature vector, then learn the parameters instead. Although assumptions from the priori knowledge can potentially explore an appropriate feature vector, selecting a biased one that insufficiently represents the system usually leads to the poor learning performance. To avoid this disadvantage, this paper introduces kernel methods to implicitly propose a learnable feature vector instead of a pre-selected one. More specifically, the feature vector is estimated from a reference set which contains all critical state-action pairs observed so far, and it can be updated by either adding a new pair or replace an existing one in the reference set. Thus the approximate Q-function keeps adjusting itself as the knowledge about the system accumulates via observations. Our algorithm is designed in both batch mode and online mode in the context of the traffic signal control. In addition, the convergence of this algorithm is experimentally supported. Furthermore, some regularization methods are proposed to avoid overfitting of Q-function on the noisy observations. Finally, A simulation on the traffic signal control in a single intersection is provided, and the performance of this algorithm is compared with Q-learning, in which the Q-function is numerically estimated for each state-action pair without approximation.
引用
收藏
页码:1277 / 1282
页数:6
相关论文
共 50 条
  • [1] Novel Feature Selection and Kernel-Based Value Approximation Method for Reinforcement Learning
    Jakab, Hunor Sandor
    Csato, Lehel
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2013, 2013, 8131 : 170 - 177
  • [2] Kernel-Based Reinforcement Learning
    Hu, Guanghua
    Qiu, Yuqin
    Xiang, Liming
    INTELLIGENT COMPUTING, PART I: INTERNATIONAL CONFERENCE ON INTELLIGENT COMPUTING, ICIC 2006, PART I, 2006, 4113 : 757 - 766
  • [3] Kernel-Based Reinforcement Learning
    Dirk Ormoneit
    Śaunak Sen
    Machine Learning, 2002, 49 : 161 - 178
  • [4] Kernel-based reinforcement learning
    Ormoneit, D
    Sen, S
    MACHINE LEARNING, 2002, 49 (2-3) : 161 - 178
  • [5] Kernel-based learning and feature selection analysis for cancer diagnosis
    Medjahed, Seyyid Ahmed
    Saadi, Tamazouzt Ait
    Benyettou, Abdelkader
    Ouali, Mohammed
    APPLIED SOFT COMPUTING, 2017, 51 : 39 - 48
  • [6] Adaptive Traffic Signal Control Method Based on Offline Reinforcement Learning
    Wang, Lei
    Wang, Yu-Xuan
    Li, Jian-Kang
    Liu, Yi
    Pi, Jia-Tian
    APPLIED SCIENCES-BASEL, 2024, 14 (22):
  • [7] A Reinforcement Learning Based Adaptive Traffic Signal Control for Vehicular Networks
    Krishnendhu, S. P.
    Reddy, Mainampati Vigneshwari
    Basumatary, Thulunga
    Mohandas, Prabu
    PERVASIVE COMPUTING AND SOCIAL NETWORKING, ICPCSN 2022, 2023, 475 : 547 - 561
  • [8] Adaptive Traffic Signal Control for Urban Corridor Based on Reinforcement Learning
    Liu, Lishan
    Zhuang, Xiya
    Li, Qiang
    COMPUTATIONAL AND EXPERIMENTAL SIMULATIONS IN ENGINEERING, ICCES 2024-VOL 2, 2025, 173 : 25 - 35
  • [9] Reinforcement learning for True Adaptive traffic signal control
    Abdulhai, B
    Pringle, R
    Karakoulas, GJ
    JOURNAL OF TRANSPORTATION ENGINEERING, 2003, 129 (03) : 278 - 285
  • [10] Practical Kernel-Based Reinforcement Learning
    Barreto, Andre M. S.
    Precup, Doina
    Pineau, Joelle
    JOURNAL OF MACHINE LEARNING RESEARCH, 2016, 17