Hierarchical reinforcement learning based on macro actions

被引:0
|
作者
Hao Jiang [1 ]
Gongju Wang [2 ]
Shengze Li [1 ]
Jieyuan Zhang [1 ]
Long Yan [2 ]
Xinhai Xu [1 ]
机构
[1] Chinese Academy of Military Science,Data Intelligence Division
[2] China Unicom Digital Technology Co,undefined
关键词
Hierarchical reinforcement learning; Macro action mapping model; Combat and non-combat macro actions; Rule-based execution logic;
D O I
10.1007/s40747-025-01895-9
中图分类号
学科分类号
摘要
The large action space is a key challenge in reinforcement learning. Although hierarchical methods have been proven to be effective in addressing this issue, they are not fully explored. This paper combines domain knowledge with hierarchical concepts to propose a novel Hierarchical Reinforcement Learning framework based on macro actions (HRL-MA). This framework includes a macro action mapping model that abstracts sequences of micro actions into macro actions, thereby simplifying the decision-making process. Macro actions are divided into two categories: combat macro actions (CMA) and non-combat macro actions (NO-CMA). NO-CMA are driven by decision tree-based logical rules and provide conditions for the execution of CMA. CMA form the action space of the reinforcement learning algorithm, which dynamically selects actions based on the current state. Comprehensive tests on the StarCraft II maps Simple64 and AbyssalReefLE demonstrate that the HRL-MA framework exhibits superior performance, achieving higher win rates compared to baseline algorithms. Furthermore, in mini-game scenarios, HRL-MA consistently outperforms baseline algorithms in terms of reward scores. The findings highlight the effectiveness of integrating hierarchical structures and macro actions in reinforcement learning to manage complex decision-making tasks in environments with large action spaces.
引用
收藏
相关论文
共 50 条
  • [1] Learning macro-actions in reinforcement learning
    Randlov, J
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 11, 1999, 11 : 1045 - 1051
  • [2] Reusability and Transferability of Macro Actions for Reinforcement Learning
    Chang Y.-H.
    Chang K.-Y.
    Kuo H.
    Lee C.-Y.
    ACM Transactions on Evolutionary Learning and Optimization, 2022, 2 (01):
  • [3] Composing Synergistic Macro Actions for Reinforcement Learning Agents
    Chen, Yu-Ming
    Chang, Kaun-Yu
    Liu, Chien
    Hsiao, Tsu-Ching
    Hong, Zhang-Wei
    Lee, Chun-Yi
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (05) : 7251 - 7258
  • [4] An acquiring method of macro-actions in reinforcement learning
    Yoshikawa, Takeshi
    Kurihara, Masahito
    2006 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS, VOLS 1-6, PROCEEDINGS, 2006, : 4813 - +
  • [5] A memory-based reinforcement learning model utilizing macro-actions
    Murata, M
    Ozawa, S
    ADAPTIVE AND NATURAL COMPUTING ALGORITHMS, 2005, : 78 - 81
  • [6] Automatic construction and evaluation of macro-actions in reinforcement learning
    Farahani, Marzieh Davoodabadi
    Mozayani, Nasser
    APPLIED SOFT COMPUTING, 2019, 82
  • [7] Hierarchical reinforcement learning for chip-macro placement in integrated circuit
    Tan, Zhentao
    Mu, Yadong
    PATTERN RECOGNITION LETTERS, 2024, 179 : 108 - 114
  • [8] Deep Reinforcement Learning for Decentralized Multi-Robot Exploration With Macro Actions
    Tan, Aaron Hao
    Bejarano, Federico Pizarro
    Zhu, Yuhan
    Ren, Richard
    Nejat, Goldie
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (01) : 272 - 279
  • [9] Multi-Agent/Robot Deep Reinforcement Learning with Macro-Actions
    Xiao, Yuchen
    Hoffman, Joshua
    Xia, Tian
    Amato, Christopher
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 13965 - 13966
  • [10] Automatic generation of macro-actions using genetic algorithm for reinforcement learning
    Tateyama, T
    Kawata, S
    Oguchi, T
    SICE 2002: PROCEEDINGS OF THE 41ST SICE ANNUAL CONFERENCE, VOLS 1-5, 2002, : 286 - 289