Tunnel ventilation control via an actor-critic algorithm employing nonparametric policy gradients

被引:0
|
作者
Division of Mechanical Engineering, Korea University, Seoul, 136-701, Korea, Republic of [1 ]
不详 [2 ]
机构
来源
J. Mech. Sci. Technol. | 2009年 / 2卷 / 311-323期
关键词
Ventilation - Automobile drivers - Pollution - Learning algorithms;
D O I
暂无
中图分类号
学科分类号
摘要
The appropriate operation of a tunnel ventilation system provides drivers passing through the tunnel with comfortable and safe driving conditions. Tunnel ventilation involves maintaining CO pollutant concentration and VI (visibility index) under an adequate level with operating highly energy-consuming facilities such as jet-fans. Therefore, it is significant to have an efficient operating algorithm in aspects of a safe driving environment as well as saving energy. In this research, a reinforcement learning (RL) method based on the actor-critic architecture and nonparametric policy gradients is applied as the control algorithm. The two objectives listed above, maintaining an adequate level of pollutants and minimizing power consumption, are included into a reward formulation that is a performance index to be maximized in the RL methodology. In this paper, a nonparametric approach is adopted as a promising route to perform a rigorous gradient search in a function space of policies to improve the efficacy of the actor module. Extensive simulation studies performed with real data collected from an existing tunnel system confirm that with the suggested algorithm, the control purposes were well accomplished and improved when compared to a previously developed RL-based control algorithm. © KSME & Springer 2009.
引用
收藏
相关论文
共 50 条
  • [31] Actor-Critic Learning Control With Regularization and Feature Selection in Policy Gradient Estimation
    Li, Luntong
    Li, Dazi
    Song, Tianheng
    Xu, Xin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (03) : 1217 - 1227
  • [32] Optimal Actor-Critic Policy With Optimized Training Datasets
    Banerjee, Chayan
    Chen, Zhiyong
    Noman, Nasimul
    Zamani, Mohsen
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2022, 6 (06): : 1324 - 1334
  • [33] Policy-Gradient Based Actor-Critic Algorithms
    Awate, Yogesh P.
    PROCEEDINGS OF THE 2009 WRI GLOBAL CONGRESS ON INTELLIGENT SYSTEMS, VOL III, 2009, : 505 - 509
  • [34] Meta attention for Off-Policy Actor-Critic
    Huang, Jiateng
    Huang, Wanrong
    Lan, Long
    Wu, Dan
    NEURAL NETWORKS, 2023, 163 : 86 - 96
  • [35] Exploring Policy Diversity in Parallel Actor-Critic Learning
    Zhang, Yanqiang
    Zhai, Yuanzhao
    Zhou, Gongqian
    Ding, Bo
    Feng, Dawei
    Liu, Songwang
    2022 IEEE 34TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, ICTAI, 2022, : 1196 - 1203
  • [36] Characterizing the Gap Between Actor-Critic and Policy Gradient
    Wen, Junfeng
    Kumar, Saurabh
    Gummadi, Ramki
    Schuurmans, Dale
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [37] An Actor-Critic Algorithm for the Stochastic Cutting Stock Problem
    Su, Jie-Ying
    Kang, Jia-Lin
    Jang, Shi-Shang
    PROCESSES, 2023, 11 (04)
  • [38] An actor-critic algorithm for constrained Markov decision processes
    Borkar, VS
    SYSTEMS & CONTROL LETTERS, 2005, 54 (03) : 207 - 213
  • [39] Finite-Sample Analysis of Off-Policy Natural Actor-Critic Algorithm
    Khodadadian, Sajad
    Chen, Zaiwei
    Maguluri, Siva Theja
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [40] An inertia wheel pendulum control method based on actor-critic learning algorithm
    Liu Huanlong
    Wang Zhengjie
    Jiang Bin
    Peng Hongyu
    2021 IEEE 20TH INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS (TRUSTCOM 2021), 2021, : 1281 - 1285