GRIMGEP: Learning Progress for Robust Goal Sampling in Visual Deep Reinforcement Learning

被引:2
|
作者
Kovac, Grgur [1 ]
Laversanne-Finot, Adrien [1 ]
Oudeyer, Pierre-Yves [1 ]
机构
[1] INRIA Bordeaux, Flowers Lab, F-33400 Talence, France
关键词
Goal exploration; learning progress; reinforcement learning (RL); INTRINSIC MOTIVATION; EXPLORATION; SYSTEMS;
D O I
10.1109/TCDS.2022.3216911
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Autotelic reinforcement learning (RL) agents sample their own goals, and try to reach them. They often prioritize goal sampling according to some intrinsic reward, ex. novelty or absolute learning progress (ALPs). Novelty-based approaches work robustly in unsupervised image-based environments when there are no distractors. However, they construct simple curricula that do not take the agent's performance into account: in complex environments, they often get attracted by impossible tasks. ALP-based approaches, which are often combined with a clustering mechanism, construct complex curricula tuned to the agent's current capabilities. Such curricula sample goals on which the agent is currently learning the most, and do not get attracted by impossible tasks. However, ALP approaches have not so far been applied to DRL agents perceiving complex environments directly in the image space. Goal regions guided intrinsically motivated goal exploration process (GRIMGEP), without using any expert knowledge, combines the ALP clustering approaches with novelty-based approaches and extends them to those complex scenarios. We experiment on a rich 3-D image-based environment with distractors using novelty-based exploration approaches: Skewfit and CountBased. We show that wrapping them with GRIMGEP-using them only in the cluster sampled by ALP-creates a better curriculum. The wrapped approaches are attracted less by the distractors, and achieve drastically better performances.
引用
收藏
页码:1396 / 1407
页数:12
相关论文
共 50 条
  • [1] Deep Reinforcement Learning for Robust Goal-Based Wealth Management
    Bauman, Tessa
    Gasperov, Bruno
    Begusic, Stjepan
    Kostanjcar, Zvonko
    ARTIFICIAL INTELLIGENCE APPLICATIONS AND INNOVATIONS, AIAI 2023, PT I, 2023, 675 : 69 - 80
  • [2] Goal Misgeneralization in Deep Reinforcement Learning
    Langosco, Lauro
    Koch, Jack
    Sharkey, Lee
    Pfau, Jacob
    Krueger, David
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [3] Goal Modelling for Deep Reinforcement Learning Agents
    Leung, Jonathan
    Shen, Zhiqi
    Zeng, Zhiwei
    Miao, Chunyan
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, 2021, 12975 : 271 - 286
  • [4] Tracker-Level Decision by Deep Reinforcement Learning for Robust Visual Tracking
    Huang, Wenju
    Wu, Yuwei
    Jia, Yunde
    IMAGE AND GRAPHICS, ICIG 2019, PT I, 2019, 11901 : 442 - 453
  • [5] Automation of noise sampling in deep reinforcement learning
    Karda, Kunal
    Dubey, Namit
    Kanungo, Abhas
    Gupta, Varun
    INTERNATIONAL JOURNAL OF APPLIED PATTERN RECOGNITION, 2022, 7 (01) : 15 - 23
  • [6] Curriculum goal masking for continuous deep reinforcement learning
    Eppe, Manfred
    Magg, Sven
    Wermter, Stefan
    2019 JOINT IEEE 9TH INTERNATIONAL CONFERENCE ON DEVELOPMENT AND LEARNING AND EPIGENETIC ROBOTICS (ICDL-EPIROB), 2019, : 183 - 188
  • [7] Comparison of multiple reinforcement learning and deep reinforcement learning methods for the task aimed at achieving the goal
    Parak R.
    Matousek R.
    Mendel, 2021, 27 (01) : 1 - 8
  • [8] GOATS: Goal Sampling Adaptation for Scooping with Curriculum Reinforcement Learning
    Niu, Yaru
    Jin, Shiyu
    Zhang, Zeqing
    Zhu, Jiacheng
    Zhao, Ding
    Zhang, Liangjun
    2023 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, IROS, 2023, : 1023 - 1030
  • [9] RESEARCH PROGRESS ABOUT DEEP REINFORCEMENT LEARNING
    Liu, Liu
    Chen, Lin-hui
    MECHATRONIC SYSTEMS AND CONTROL, 2023, 51 (04): : 210 - 217
  • [10] Deep Reinforcement Learning for Economics: Progress and Challenges
    Green, Etan A.
    Plunkett, E. Barry
    ACM SIGECOM EXCHANGES, 2023, 21 (01) : 49 - 53