Towards optimal control of air handling units using deep reinforcement learning and recurrent neural network

被引:115
|
作者
Zou, Zhengbo [1 ]
Yu, Xinran [1 ]
Ergan, Semiha [1 ]
机构
[1] NYU, Dept Civil & Urban Engn, MetroTech Ctr 15, Brooklyn, NY 11201 USA
关键词
HVAC control; Energy consumption; Thermal comfort; Deep reinforcement learning; Long-short-term-memory network; ENERGY-CONSUMPTION; BUILDINGS; SYSTEMS; COMFORT; MODELS;
D O I
10.1016/j.buildenv.2019.106535
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
Optimal control of heating, ventilation and air conditioning systems (HVACs) aims to minimize the energy consumption of equipment while maintaining the thermal comfort of occupants. Traditional rule-based control methods are not optimized for HVAC systems with continuous sensor readings and actuator controls. Recent developments in deep reinforcement learning (DRL) enabled control of HVACs with continuous sensor inputs and actions, while eliminating the need of building complex thermodynamic models. DRL control includes an environment, which approximates real-world HVAC operations; and an agent, that aims to achieve optimal control over the HVAC. Existing DRL control frameworks use simulation tools (e.g., EnergyPlus) to build DRL training environments with HVAC systems information, but oversimplify building geometrics. This study proposes a framework aiming to achieve optimal control over Air Handling Units (AHUs) by implementing longshort-term-memory (LSTM) networks to approximate real-world HVAC operations to build DRL training environments. The framework also implements state-of-the-art DRL algorithms (e.g., deep deterministic policy gradient) for optimal control over the AHUs. Three AHUs, each with two-years of building automation system (BAS) data, were used as testbeds for evaluation. Our LSTM-based DRL training environments, built using the first year's BAS data, achieved an average mean square error of 0.0015 across 16 normalized AHU parameters. When deployed in the testing environments, which were built using the second year's BAS data of the same AHUs, the DRL agents achieved 27%-30% energy saving comparing to the actual energy consumption, while maintaining the predicted percentage of discomfort (PPD) at 10%.
引用
收藏
页数:15
相关论文
共 50 条
  • [41] Dynamic Vehicle Traffic Control Using Deep Reinforcement Learning in Automated Material Handling System
    Kang, Younkook
    Lyu, Sungwon
    Kim, Jeeyung
    Park, Bongjoon
    Cho, Sungzoon
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 9949 - 9950
  • [42] Robust reinforcement learning control using integral quadratic constraints for recurrent neural networks
    Anderson, Charles W.
    Young, Peter Michael
    Buehner, Michael R.
    Knight, James N.
    Bush, Keith A.
    Hittle, Douglas C.
    IEEE TRANSACTIONS ON NEURAL NETWORKS, 2007, 18 (04): : 993 - 1002
  • [43] Application of Deep Reinforcement Learning in Optimal Operation of Distribution Network
    Hu W.
    Cao D.
    Huang Q.
    Zhang B.
    Li S.
    Chen Z.
    Dianli Xitong Zidonghua/Automation of Electric Power Systems, 2023, 47 (14): : 174 - 191
  • [44] Optimization of Air Network Resources Based on Deep Reinforcement Learning
    Zhang, Yuanwei
    Cui, Haixia
    2024 4TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND INTELLIGENT SYSTEMS ENGINEERING, MLISE 2024, 2024, : 145 - 151
  • [45] Optimal Terminal Box Control Algorithms for Single Duct Air Handling Units
    Cho, Young-Hum
    Liu, Mingsheng
    ES2008: PROCEEDINGS OF THE 2ND INTERNATIONAL CONFERENCE ON ENERGY SUSTAINABILITY, VOL 2, 2009, : 203 - 208
  • [46] Intelligent fault diagnosis for air handing units based on improved generative adversarial network and deep reinforcement learning
    Yan, Ke
    Lu, Cheng
    Ma, Xiang
    Ji, Zhiwei
    Huang, Jing
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 240
  • [47] A Reinforcement Learning Neural Network for Robotic Manipulator Control
    Hu, Yazhou
    Si, Bailu
    NEURAL COMPUTATION, 2018, 30 (07) : 1983 - 2004
  • [48] Production Scheduling based on Deep Reinforcement Learning using Graph Convolutional Neural Network
    Seito, Takanari
    Munakata, Satoshi
    ICAART: PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE, VOL 2, 2020, : 766 - 772
  • [49] Multiobjective Reinforcement Learning for Cognitive Satellite Communications Using Deep Neural Network Ensembles
    Rodrigues Ferreira, Paulo Victor
    Paffenroth, Randy
    Wyglinski, Alexander M.
    Hackett, Timothy M.
    Bilen, Sven G.
    Reinhart, Richard C.
    Mortensen, Dale J.
    IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2018, 36 (05) : 1030 - 1041
  • [50] Optimal deep learning neural network using ISSA for diagnosing the oral cancer
    Huang, Qirui
    Ding, Huan
    Razmjooy, Navid
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2023, 84