Bayesian Strategy Networks Based Soft Actor-Critic Learning

被引:0
|
作者
Yang, Qin [1 ]
Parasuraman, Ramviyas [2 ]
机构
[1] Bradley Univ, Bradley Hall 195,1501 Bradley Ave, Peoria, IL 61625 USA
[2] Univ Georgia, 415 Boyd Res & Educ Ctr, Athens, GA 30602 USA
关键词
Strategy; bayesian networks; deep reinforcement learning; soft actor-critic; utility; expectation; REINFORCEMENT; LEVEL;
D O I
10.1145/3643862
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Astrategy refers to the rules that the agent chooses the available actions to achieve goals. Adopting reasonable strategies is challenging but crucial for an intelligent agent with limited resources working in hazardous, unstructured, and dynamic environments to improve the system's utility, decrease the overall cost, and increase mission success probability. This article proposes a novel hierarchical strategy decomposition approach based on Bayesian chaining to separate an intricate policy into several simple sub-policies and organize their relationships as Bayesian strategy networks (BSN). We integrate this approach into the state-of-the-art DRL method-soft actor-critic (SAC), and build the corresponding Bayesian soft actor-critic (BSAC) model by organizing several sub-policies as a joint policy. Our method achieves the state-of-the-art performance on the standard continuous control benchmarks in the OpenAI Gym environment. The results demonstrate that the promising potential of the BSAC method significantly improves training efficiency. Furthermore, we extend the topic to the Multi-Agent systems (MAS), discussing the potential research fields and directions.
引用
收藏
页数:24
相关论文
共 50 条
  • [31] A soft actor-critic reinforcement learning algorithm for network intrusion detection
    Li, Zhengfa
    Huang, Chuanhe
    Deng, Shuhua
    Qiu, Wanyu
    Gao, Xieping
    COMPUTERS & SECURITY, 2023, 135
  • [32] A Novel Actor-Critic Motor Reinforcement Learning for Continuum Soft Robots
    Pantoja-Garcia, Luis
    Parra-Vega, Vicente
    Garcia-Rodriguez, Rodolfo
    Vazquez-Garcia, Carlos Ernesto
    ROBOTICS, 2023, 12 (05)
  • [33] Improving Generalization of Reinforcement Learning with Minimax Distributional Soft Actor-Critic
    Ren, Yangang
    Duan, Jingliang
    Li, Shengbo Eben
    Guan, Yang
    Sun, Qi
    2020 IEEE 23RD INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2020,
  • [34] A Soft Actor-Critic Algorithm for Sequential Recommendation
    Hong, Hyejin
    Kimurn, Yusuke
    Hatano, Kenji
    DATABASE AND EXPERT SYSTEMS APPLICATIONS, PT I, DEXA 2024, 2024, 14910 : 258 - 266
  • [35] SOFT ACTOR-CRITIC ALGORITHM WITH ADAPTIVE NORMALIZATION
    Gao, Xiaonan
    Wu, Ziyi
    Zhu, Xianchao
    Cai, Lei
    JOURNAL OF NONLINEAR FUNCTIONAL ANALYSIS, 2025, 2025
  • [36] Granular computing in actor-critic learning
    Peters, James F.
    2007 IEEE SYMPOSIUM ON FOUNDATIONS OF COMPUTATIONAL INTELLIGENCE, VOLS 1 AND 2, 2007, : 59 - 64
  • [37] Peg-Hole Robotic Disassembly Compliant Strategy based on Soft Actor-Critic Algorithm
    Zhang, Xiaolong
    Liu, Jiayi
    Qi, Lei
    Xu, Wenjun
    Peng, Xianxu
    Wu, Tianshu
    2024 IEEE 20TH INTERNATIONAL CONFERENCE ON AUTOMATION SCIENCE AND ENGINEERING, CASE 2024, 2024, : 528 - 533
  • [38] Actor-Critic Learning Based QoS-Aware Scheduler for Reconfigurable Wireless Networks
    Mollahasani, Shahram
    Erol-Kantarci, Melike
    Hirab, Mahdi
    Dehghan, Hoda
    Wilson, Rodney
    IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2022, 9 (01): : 45 - 54
  • [39] Proactive Content Caching Based on Actor-Critic Reinforcement Learning for Mobile Edge Networks
    Jiang, Wei
    Feng, Daquan
    Sun, Yao
    Feng, Gang
    Wang, Zhenzhong
    Xia, Xiang-Gen
    IEEE TRANSACTIONS ON COGNITIVE COMMUNICATIONS AND NETWORKING, 2022, 8 (02) : 1239 - 1252
  • [40] Actor-critic learning-based energy optimization for UAV access and backhaul networks
    Yaxiong Yuan
    Lei Lei
    Thang X. Vu
    Symeon Chatzinotas
    Sumei Sun
    Björn Ottersten
    EURASIP Journal on Wireless Communications and Networking, 2021