Disentangling Controlled Effects for Hierarchical Reinforcement Learning

被引:0
|
作者
Corcoll, Oriol [1 ]
Vicente, Raul [1 ]
机构
[1] Univ Tartu, Inst Comp Sci, Tartu, Estonia
关键词
unsupervised reinforcement learning; reinforcement learning; causality;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Exploration and credit assignment are still challenging problems for RL agents under sparse rewards. We argue that these challenges arise partly due to the intrinsic rigidity of operating at the level of actions. Actions can precisely define how to perform an activity but are ill-suited to describe what activity to perform. Instead, controlled effects describe transformations in the environment caused by the agent. These transformations are inherently composable and temporally abstract, making them ideal for descriptive tasks. This work introduces CEHRL1, a hierarchical method leveraging the compositional nature of controlled effects to expedite the learning of task-specific behavior and aid exploration. Borrowing counterfactual and normality measures from causal literature, CEHRL learns an implicit hierarchy of transformations an agent can perform on the environment. This hierarchy allows a high-level policy to set temporally abstract goals and, by doing so, long-horizon credit assignment. Experimental results show that using effects instead of actions provides a more efficient exploration mechanism. Moreover, by leveraging prior knowledge in the hierarchy, CEHRL assigns credit to few effects instead of many actions and consequently learns tasks more rapidly.
引用
收藏
页数:23
相关论文
共 50 条
  • [1] Disentangling Transfer in Continual Reinforcement Learning
    Wolczyk, Maciej
    Zajac, Michal
    Pascanu, Razvan
    Kucinski, Lukasz
    Milos, Piotr
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [2] DISENTANGLING EFFECTS OF DELAY OF REINFORCEMENT FROM RATE OF REINFORCEMENT
    WASSERMAN, EA
    HUSSAR, KA
    BHATT, RS
    BULLETIN OF THE PSYCHONOMIC SOCIETY, 1985, 23 (04) : 283 - 283
  • [3] Deep Adversarial Reinforcement Learning for Object Disentangling
    Laux, Melvin
    Arenz, Oleg
    Peters, Jan
    Pajarinen, Joni
    2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 5504 - 5510
  • [4] Hierarchical disentangling network for object representation learning
    Qiao, Shishi
    Wang, Ruiping
    Shan, Shiguang
    Chen, Xilin
    PATTERN RECOGNITION, 2023, 140
  • [5] Concurrent Hierarchical Reinforcement Learning
    Marthi, Bhaskara
    Russell, Stuart
    Latham, David
    Guestrin, Carlos
    19TH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI-05), 2005, : 779 - 785
  • [6] Hierarchical reinforcement learning with OMQ
    Shen, Jing
    Liu, Haibo
    Gu, Guochang
    PROCEEDINGS OF THE FIFTH IEEE INTERNATIONAL CONFERENCE ON COGNITIVE INFORMATICS, VOLS 1 AND 2, 2006, : 584 - 588
  • [7] Hierarchical Imitation and Reinforcement Learning
    Le, Hoang M.
    Jiang, Nan
    Agarwal, Alekh
    Dudik, Miroslav
    Yue, Yisong
    Daume, Hal, III
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [8] On Efficiency in Hierarchical Reinforcement Learning
    Wen, Zheng
    Precup, Doina
    Ibrahimi, Morteza
    Barreto, Andre
    Van Roy, Benjamin
    Singh, Satinder
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [9] Budgeted Hierarchical Reinforcement Learning
    Leon, Aurelia
    Denoyer, Ludovic
    2018 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2018,
  • [10] Hierarchical Reinforcement Learning: A Comprehensive Survey
    Pateria, Shubham
    Subagdja, Budhitama
    Tan, Ah-hwee
    Quek, Chai
    ACM COMPUTING SURVEYS, 2021, 54 (05)