Disentangling Controlled Effects for Hierarchical Reinforcement Learning

被引:0
|
作者
Corcoll, Oriol [1 ]
Vicente, Raul [1 ]
机构
[1] Univ Tartu, Inst Comp Sci, Tartu, Estonia
关键词
unsupervised reinforcement learning; reinforcement learning; causality;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Exploration and credit assignment are still challenging problems for RL agents under sparse rewards. We argue that these challenges arise partly due to the intrinsic rigidity of operating at the level of actions. Actions can precisely define how to perform an activity but are ill-suited to describe what activity to perform. Instead, controlled effects describe transformations in the environment caused by the agent. These transformations are inherently composable and temporally abstract, making them ideal for descriptive tasks. This work introduces CEHRL1, a hierarchical method leveraging the compositional nature of controlled effects to expedite the learning of task-specific behavior and aid exploration. Borrowing counterfactual and normality measures from causal literature, CEHRL learns an implicit hierarchy of transformations an agent can perform on the environment. This hierarchy allows a high-level policy to set temporally abstract goals and, by doing so, long-horizon credit assignment. Experimental results show that using effects instead of actions provides a more efficient exploration mechanism. Moreover, by leveraging prior knowledge in the hierarchy, CEHRL assigns credit to few effects instead of many actions and consequently learns tasks more rapidly.
引用
收藏
页数:23
相关论文
共 50 条
  • [31] Hierarchical Reinforcement Learning for Integrated Recommendation
    Xie, Ruobing
    Zhang, Shaoliang
    Wang, Rui
    Xia, Feng
    Lin, Leyu
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 4521 - 4528
  • [32] Evaluating skills in hierarchical reinforcement learning
    Farahani, Marzieh Davoodabadi
    Mozayani, Nasser
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2020, 11 (10) : 2407 - 2420
  • [33] Hierarchical average reward reinforcement learning
    Department of Computing Science, University of Alberta, Edmonton, Alta. T6G 2E8, Canada
    不详
    Journal of Machine Learning Research, 2007, 8 : 2629 - 2669
  • [34] Autonomous Reinforcement Learning with Hierarchical REPS
    Daniel, Christian
    Neumann, Gerhard
    Peters, Jan
    2013 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2013,
  • [35] Scalable Evolutionary Hierarchical Reinforcement Learning
    Abramowitz, Sasha
    Nitschke, Geoff
    PROCEEDINGS OF THE 2022 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE COMPANION, GECCO 2022, 2022, : 272 - 275
  • [36] A Neural Signature of Hierarchical Reinforcement Learning
    Ribas-Fernandes, Jose J. F.
    Solway, Alec
    Diuk, Carlos
    McGuire, Joseph T.
    Barto, Andrew G.
    Niv, Yael
    Botvinick, Matthew M.
    NEURON, 2011, 71 (02) : 370 - 379
  • [37] Interaction Pattern Disentangling for Multi-Agent Reinforcement Learning
    Liu, Shunyu
    Song, Jie
    Zhou, Yihe
    Yu, Na
    Chen, Kaixuan
    Feng, Zunlei
    Song, Mingli
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (12) : 8157 - 8172
  • [38] Optimal Hierarchical Learning Path Design With Reinforcement Learning
    Li, Xiao
    Xu, Hanchen
    Zhang, Jinming
    Chang, Hua-hua
    APPLIED PSYCHOLOGICAL MEASUREMENT, 2021, 45 (01) : 54 - 70
  • [39] Learning Generalizable Locomotion Skills with Hierarchical Reinforcement Learning
    Li, Tianyu
    Lambert, Nathan
    Calandra, Roberto
    Meier, Franziska
    Rai, Akshara
    2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 413 - 419
  • [40] Effectively Learning Initiation Sets in Hierarchical Reinforcement Learning
    Bagaria, Akhil
    Abbatematteo, Ben
    Gottesman, Omer
    Corsaro, Matt
    Rammohan, Sreehari
    Konidaris, George
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,