Disentangling Controlled Effects for Hierarchical Reinforcement Learning

被引:0
|
作者
Corcoll, Oriol [1 ]
Vicente, Raul [1 ]
机构
[1] Univ Tartu, Inst Comp Sci, Tartu, Estonia
关键词
unsupervised reinforcement learning; reinforcement learning; causality;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Exploration and credit assignment are still challenging problems for RL agents under sparse rewards. We argue that these challenges arise partly due to the intrinsic rigidity of operating at the level of actions. Actions can precisely define how to perform an activity but are ill-suited to describe what activity to perform. Instead, controlled effects describe transformations in the environment caused by the agent. These transformations are inherently composable and temporally abstract, making them ideal for descriptive tasks. This work introduces CEHRL1, a hierarchical method leveraging the compositional nature of controlled effects to expedite the learning of task-specific behavior and aid exploration. Borrowing counterfactual and normality measures from causal literature, CEHRL learns an implicit hierarchy of transformations an agent can perform on the environment. This hierarchy allows a high-level policy to set temporally abstract goals and, by doing so, long-horizon credit assignment. Experimental results show that using effects instead of actions provides a more efficient exploration mechanism. Moreover, by leveraging prior knowledge in the hierarchy, CEHRL assigns credit to few effects instead of many actions and consequently learns tasks more rapidly.
引用
收藏
页数:23
相关论文
共 50 条
  • [41] Hierarchical Reinforcement Learning Under Mixed Observability
    Hai Nguyen
    Yang, Zhihan
    Baisero, Andrea
    Ma, Xiao
    Platt, Robert
    Amato, Christopher
    ALGORITHMIC FOUNDATIONS OF ROBOTICS XV, 2023, 25 : 188 - 204
  • [42] Hierarchical reinforcement learning model for military simulations
    Sidhu, Amandeep Singh
    Chaudhari, Narendra S.
    Goh, Ghee Ming
    2006 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORK PROCEEDINGS, VOLS 1-10, 2006, : 2572 - +
  • [43] Hierarchical multi-agent reinforcement learning
    Mohammad Ghavamzadeh
    Sridhar Mahadevan
    Rajbala Makar
    Autonomous Agents and Multi-Agent Systems, 2006, 13 : 197 - 229
  • [44] Hierarchical reinforcement learning guidance with threat avoidance
    Li Bohao
    Wu Yunjie
    Li Guofei
    JOURNAL OF SYSTEMS ENGINEERING AND ELECTRONICS, 2022, 33 (05) : 1173 - 1185
  • [45] Hierarchical Reinforcement Learning with Clustering Abstract Machines
    Alexey, Skrynnik
    Panov, Aleksandr, I
    ARTIFICIAL INTELLIGENCE: (RCAI 2019), 2019, 1093 : 30 - 43
  • [46] HIERARCHICAL KICKSTARTING FOR SKILL TRANSFER IN REINFORCEMENT LEARNING
    Matthews, Michael
    Samvelyan, Mikayel
    Parker-Holder, Jack
    Grefenstette, Edward
    Rocktaschel, Tim
    CONFERENCE ON LIFELONG LEARNING AGENTS, VOL 199, 2022, 199
  • [47] Emotion-based hierarchical reinforcement learning
    Zhou, WD
    Coggins, R
    DESIGN AND APPLICATION OF HYBRID INTELLIGENT SYSTEMS, 2003, 104 : 951 - 960
  • [48] HLifeRL: A hierarchical lifelong reinforcement learning framework
    Ding, Fan
    Zhu, Fei
    JOURNAL OF KING SAUD UNIVERSITY-COMPUTER AND INFORMATION SCIENCES, 2022, 34 (07) : 4312 - 4321
  • [49] Planning-Augmented Hierarchical Reinforcement Learning
    Gieselmann, Robert
    Pokorny, Florian T.
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (03) : 5097 - 5104
  • [50] Hierarchical reinforcement learning using path clustering
    Gil, Paulo
    Nunes, Luis
    PROCEEDINGS OF THE 2013 8TH IBERIAN CONFERENCE ON INFORMATION SYSTEMS AND TECHNOLOGIES (CISTI 2013), 2013,