Real-world challenges for multi-agent reinforcement learning in grid-interactive buildings

被引:21
|
作者
Nweye, Kingsley [1 ]
Liu, Bo [2 ]
Stone, Peter [2 ]
Nagy, Zoltan [1 ]
机构
[1] Univ Texas Austin, Dept Civil Architectural & Environm Engn, Intelligent Environm Lab, 301 E Dean Keeton St Stop,ECJ 4 200, Austin, TX 78712 USA
[2] Univ Texas Austin, Dept Comp Sci, 2317 Speedway,GDC 2 302, Austin, TX 78712 USA
关键词
Grid-interactive buildings; Benchmarking; Reinforcement learning; DEMAND RESPONSE;
D O I
10.1016/j.egyai.2022.100202
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Building upon prior research that highlighted the need for standardizing environments for building control research, and inspired by recently introduced challenges for real life reinforcement learning (RL) control, here we propose a non-exhaustive set of nine real world challenges for RL control in grid-interactive buildings (GIBs). We argue that research in this area should be expressed in this framework in addition to providing a standardized environment for repeatability. Advanced controllers such as model predictive control (MPC) and RL control have both advantages and disadvantages that prevent them from being implemented in real world problems. Comparisons between the two are rare, and often biased. By focusing on the challenges, we can investigate the performance of the controllers under a variety of situations and generate a fair comparison. As a demonstration, we implement the offline learning challenge in CityLearn, an OpenAI Gym environment for the easy implementation of RL agents in a demand response setting to reshape the aggregated curve of electricity demand by controlling the energy storage of a diverse set of buildings in a district. We use CityLearn to study the impact of different levels of domain knowledge and complexity of RL algorithms and show that the sequence of operations (SOOs) utilized in a rule based controller (RBC) that provides fixed logs to RL agents during offline training affect the performance of the agents when evaluated on a set of four energy flexibility metrics. Longer offline training from an optimized RBC leads to improved performance in the long run. RL agents that train on the logs from a simplified RBC risk poorer performance as the offline training period increases. We also observe no impact on performance from information sharing amongst agents. We call for a more interdisciplinary effort of the research community to address the real world challenges, and unlock the potential of GIB controllers.
引用
收藏
页数:10
相关论文
共 50 条
  • [21] Multi-agent Techniques to Solve a Real-World Warehouse Problem
    Acs, Botond
    Dora, Laszlo
    Jakab, Oliver
    Varga, Laszlo Z.
    ADVANCES IN PRACTICAL APPLICATIONS OF AGENTS, MULTI-AGENT SYSTEMS, AND SOCIAL GOOD: THE PAAMS COLLECTION, PAAMS 2021, 2021, 12946 : 1 - 13
  • [22] Optimizations of a Multi-Agent System for a Real-World Warehouse Problem
    Ács B.
    Dóra L.
    Jakab O.
    Jüttner A.
    Madarasi P.
    Varga L.Z.
    SN Computer Science, 3 (6)
  • [23] Multi-Agent Deep Reinforcement Learning Applications in Cybersecurity: challenges and perspectives
    Tolba, Zakaria
    Dehimi, Nour El Houda
    Galland, Stephane
    Boukelloul, Soufiene
    Guassmi, Djaber
    2024 1ST INTERNATIONAL CONFERENCE ON ELECTRICAL, COMPUTER, TELECOMMUNICATION AND ENERGY TECHNOLOGIES, ECTE-TECH, 2024,
  • [24] Optimizing the Operation of Grid-Interactive Efficient Buildings (GEBs) Using Machine Learning
    Copiaco, Czarina
    Nour, Mutasim
    SUSTAINABILITY, 2024, 16 (20)
  • [25] Fuzzy Reinforcement Learning Multi-agent System for Comfort and Energy Management in Buildings
    Kofinas, Panagiotis
    Dounis, Anastasios
    Korkidis, Panagiotis
    PROCEEDINGS OF SIXTH INTERNATIONAL CONGRESS ON INFORMATION AND COMMUNICATION TECHNOLOGY (ICICT 2021), VOL 2, 2022, 236 : 291 - 310
  • [26] Enhancing Grid-Interactive Buildings Demand Response: Sequential Update-Based Multiagent Deep Reinforcement Learning Approach
    Han, Yinghua
    Wu, Jingrun
    Chen, Haoqi
    Si, Fangyuan
    Cao, Zhiao
    Zhao, Qiang
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (14): : 24439 - 24451
  • [27] Multi-Agent Cognition Difference Reinforcement Learning for Multi-Agent Cooperation
    Wang, Huimu
    Qiu, Tenghai
    Liu, Zhen
    Pu, Zhiqiang
    Yi, Jianqiang
    Yuan, Wanmai
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [28] Multi-Agent Reinforcement Learning With Distributed Targeted Multi-Agent Communication
    Xu, Chi
    Zhang, Hui
    Zhang, Ya
    2023 35TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2023, : 2915 - 2920
  • [29] Multi-Agent Uncertainty Sharing for Cooperative Multi-Agent Reinforcement Learning
    Chen, Hao
    Yang, Guangkai
    Zhang, Junge
    Yin, Qiyue
    Huang, Kaiqi
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [30] Hierarchical multi-agent reinforcement learning
    Mohammad Ghavamzadeh
    Sridhar Mahadevan
    Rajbala Makar
    Autonomous Agents and Multi-Agent Systems, 2006, 13 : 197 - 229