Function Approximation for Reinforcement Learning Controller for Energy from

被引:0
|
作者
Sarkar, Soumyendu [1 ]
Gundecha, Vineet [1 ]
Ghorbanpour, Sahand [1 ]
Shmakov, Alexander [1 ]
Babu, Ashwin Ramesh [1 ]
Naug, Avisek [1 ]
Pichard, Alexandre [2 ]
Cocho, Mathieu [2 ]
机构
[1] Hewlett Packard Enterprise, Houston, TX 77070 USA
[2] Carnegie Clean Energy, Fremantle, WA, Australia
来源
PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023 | 2023年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The industrial multi-generator Wave Energy Converters (WEC) must handle multiple simultaneous waves coming from different directions called spread waves. These complex devices in challenging circumstances need controllers with multiple objectives of energy capture efficiency, reduction of structural stress to limit maintenance, and proactive protection against high waves. The Multi-Agent Reinforcement Learning (MARL) controller trained with Proximal Policy Optimization (PPO) algorithm can handle these complexities. In this paper, we explore different function approximations for the policy and critic networks in modeling the sequential nature of the system dynamics and find that they are key to better performance. We investigated the performance of a fully connected neural network (FCN), LSTM, and Transformer model variants with varying depths and gated residual connections. Our results show that the transformer model of moderate depth with gated residual connections around the multi-head attention, multi-layer perceptron, and the transformer block (STrXL) proposed in this paper is optimal and boosts energy efficiency by an average of 22.1% for these complex spread waves over the existing spring damper (SD) controller. Furthermore, unlike the default SD controller, the transformer controller almost eliminated the mechanical stress from the rotational yaw motion for angled waves. Demo: https://tinyurl.com/yueda3jh
引用
收藏
页码:6201 / 6209
页数:9
相关论文
共 50 条
  • [1] From free energy to expected energy: Improving energy-based value function approximation in reinforcement learning
    Elfwing, Stefan
    Uchibe, Eiji
    Doya, Kenji
    NEURAL NETWORKS, 2016, 84 : 17 - 27
  • [2] Evolutionary function approximation for reinforcement learning
    Whiteson, Shimon
    Stone, Peter
    JOURNAL OF MACHINE LEARNING RESEARCH, 2006, 7 : 877 - 917
  • [3] Multiagent reinforcement learning using function approximation
    Abul, O
    Polat, F
    Alhajj, R
    IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART C-APPLICATIONS AND REVIEWS, 2000, 30 (04): : 485 - 497
  • [4] Resilient Multiagent Reinforcement Learning With Function Approximation
    Ye, Lintao
    Figura, Martin
    Lin, Yixuan
    Pal, Mainak
    Das, Pranoy
    Liu, Ji
    Gupta, Vijay
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2024, 69 (12) : 8497 - 8512
  • [5] Ensemble Methods for Reinforcement Learning with Function Approximation
    Fausser, Stefan
    Schwenker, Friedhelm
    MULTIPLE CLASSIFIER SYSTEMS, 2011, 6713 : 56 - 65
  • [6] Distributional reinforcement learning with linear function approximation
    Bellemare, Marc G.
    Le Roux, Nicolas
    Castro, Pablo Samuel
    Moitra, Subhodeep
    22ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 89, 2019, 89
  • [7] Reinforcement learning with function approximation converges to a region
    Gordon, GJ
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 13, 2001, 13 : 1040 - 1046
  • [8] Parallel reinforcement learning with linear function approximation
    Grounds, Matthew
    Kudenko, Daniel
    ADAPTIVE AGENTS AND MULTI-AGENT SYSTEMS, 2008, 4865 : 60 - 74
  • [9] Safe Reinforcement Learning with Linear Function Approximation
    Amani, Sanae
    Thrampoulidis, Christos
    Yang, Lin F.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [10] Reinforcement learning with function approximation for cooperative navigation tasks
    Melo, Francisco S.
    Ribeiro, M. Isabel
    2008 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, VOLS 1-9, 2008, : 3321 - +