Tunnel ventilation control via an actor-critic algorithm employing nonparametric policy gradients

被引:0
|
作者
Division of Mechanical Engineering, Korea University, Seoul, 136-701, Korea, Republic of [1 ]
不详 [2 ]
机构
来源
J. Mech. Sci. Technol. | 2009年 / 2卷 / 311-323期
关键词
Ventilation - Automobile drivers - Pollution - Learning algorithms;
D O I
暂无
中图分类号
学科分类号
摘要
The appropriate operation of a tunnel ventilation system provides drivers passing through the tunnel with comfortable and safe driving conditions. Tunnel ventilation involves maintaining CO pollutant concentration and VI (visibility index) under an adequate level with operating highly energy-consuming facilities such as jet-fans. Therefore, it is significant to have an efficient operating algorithm in aspects of a safe driving environment as well as saving energy. In this research, a reinforcement learning (RL) method based on the actor-critic architecture and nonparametric policy gradients is applied as the control algorithm. The two objectives listed above, maintaining an adequate level of pollutants and minimizing power consumption, are included into a reward formulation that is a performance index to be maximized in the RL methodology. In this paper, a nonparametric approach is adopted as a promising route to perform a rigorous gradient search in a function space of policies to improve the efficacy of the actor module. Extensive simulation studies performed with real data collected from an existing tunnel system confirm that with the suggested algorithm, the control purposes were well accomplished and improved when compared to a previously developed RL-based control algorithm. © KSME & Springer 2009.
引用
收藏
相关论文
共 50 条
  • [41] Network Congestion Control Algorithm Based on Actor-Critic Reinforcement Learning Model
    Xu, Tao
    Gong, Lina
    Zhang, Wei
    Li, Xuhong
    Wang, Xia
    Pan, Wenwen
    ADVANCES IN MATERIALS, MACHINERY, ELECTRONICS II, 2018, 1955
  • [42] On Finite-Time Convergence of Actor-Critic Algorithm
    Qiu S.
    Yang Z.
    Ye J.
    Wang Z.
    IEEE Journal on Selected Areas in Information Theory, 2021, 2 (02): : 652 - 664
  • [43] An Adaptive Threshold for the Canny Edge With Actor-Critic Algorithm
    Choi, Keong-Hun
    Ha, Jong-Eun
    IEEE ACCESS, 2023, 11 : 67058 - 67069
  • [44] Actor-Critic Algorithm with Maximum-Entropy Correction
    Jiang Y.-B.
    Liu Q.
    Hu Z.-H.
    Liu, Quan (quanliu@suda.edu.cn), 1897, Science Press (43): : 1897 - 1908
  • [45] Actor-Critic Algorithm for Optimal Synchronization of Kuramoto Oscillator
    Vrushabh, D.
    Shalini, K.
    Sonam, K.
    2020 7TH INTERNATIONAL CONFERENCE ON CONTROL, DECISION AND INFORMATION TECHNOLOGIES (CODIT'20), VOL 1, 2020, : 391 - 396
  • [46] Optimal Elevator Group Control via Deep Asynchronous Actor-Critic Learning
    Wei, Qinglai
    Wang, Lingxiao
    Liu, Yu
    Polycarpou, Marios M.
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (12) : 5245 - 5256
  • [47] Robust Offline Actor-Critic with On-Policy Regularized Policy Evaluation
    Cao, Shuo
    Wang, Xuesong
    Cheng, Yuhu
    IEEE-CAA JOURNAL OF AUTOMATICA SINICA, 2024, 11 (12) : 2497 - 2511
  • [48] Event-triggered receding horizon control via actor-critic design
    Lu Dong
    Xin Yuan
    Changyin Sun
    Science China Information Sciences, 2020, 63
  • [49] Event-triggered receding horizon control via actor-critic design
    Lu DONG
    Xin YUAN
    Changyin SUN
    Science China(Information Sciences), 2020, 63 (05) : 131 - 145
  • [50] Robust Offline Actor-Critic With On-policy Regularized Policy Evaluation
    Shuo Cao
    Xuesong Wang
    Yuhu Cheng
    IEEE/CAA Journal of Automatica Sinica, 2024, 11 (12) : 2497 - 2511