GRIMGEP: Learning Progress for Robust Goal Sampling in Visual Deep Reinforcement Learning

被引:2
|
作者
Kovac, Grgur [1 ]
Laversanne-Finot, Adrien [1 ]
Oudeyer, Pierre-Yves [1 ]
机构
[1] INRIA Bordeaux, Flowers Lab, F-33400 Talence, France
关键词
Goal exploration; learning progress; reinforcement learning (RL); INTRINSIC MOTIVATION; EXPLORATION; SYSTEMS;
D O I
10.1109/TCDS.2022.3216911
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Autotelic reinforcement learning (RL) agents sample their own goals, and try to reach them. They often prioritize goal sampling according to some intrinsic reward, ex. novelty or absolute learning progress (ALPs). Novelty-based approaches work robustly in unsupervised image-based environments when there are no distractors. However, they construct simple curricula that do not take the agent's performance into account: in complex environments, they often get attracted by impossible tasks. ALP-based approaches, which are often combined with a clustering mechanism, construct complex curricula tuned to the agent's current capabilities. Such curricula sample goals on which the agent is currently learning the most, and do not get attracted by impossible tasks. However, ALP approaches have not so far been applied to DRL agents perceiving complex environments directly in the image space. Goal regions guided intrinsically motivated goal exploration process (GRIMGEP), without using any expert knowledge, combines the ALP clustering approaches with novelty-based approaches and extends them to those complex scenarios. We experiment on a rich 3-D image-based environment with distractors using novelty-based exploration approaches: Skewfit and CountBased. We show that wrapping them with GRIMGEP-using them only in the cluster sampled by ALP-creates a better curriculum. The wrapped approaches are attracted less by the distractors, and achieve drastically better performances.
引用
收藏
页码:1396 / 1407
页数:12
相关论文
共 50 条
  • [31] Visual Rationalizations in Deep Reinforcement Learning for Atari Games
    Weitkamp, Laurens
    van der Pol, Elise
    Akata, Zeynep
    ARTIFICIAL INTELLIGENCE, BNAIC 2018, 2019, 1021 : 151 - 165
  • [32] Deep Reinforcement Learning with Iterative Shift for Visual Tracking
    Ren, Liangliang
    Yuan, Xin
    Lu, Jiwen
    Yang, Ming
    Zhou, Jie
    COMPUTER VISION - ECCV 2018, PT IX, 2018, 11213 : 697 - 713
  • [33] Visual Tracking via Hierarchical Deep Reinforcement Learning
    Zhang, Dawei
    Zheng, Zhonglong
    Jia, Riheng
    Li, Minglu
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 3315 - 3323
  • [34] On the Use of Deep Reinforcement Learning for Visual Tracking: A Survey
    Cruciata, Giorgio
    Lo Presti, Liliana
    La Cascia, Marco
    IEEE ACCESS, 2021, 9 : 120880 - 120900
  • [35] Deep Reinforcement Learning for Visual Semantic Navigation with Memory
    de Andrade Santos, Iury Batista
    Romero, Roseli A. F.
    2020 XVIII LATIN AMERICAN ROBOTICS SYMPOSIUM, 2020 XII BRAZILIAN SYMPOSIUM ON ROBOTICS AND 2020 XI WORKSHOP OF ROBOTICS IN EDUCATION (LARS-SBR-WRE 2020), 2020, : 114 - 119
  • [36] Deep reinforcement learning for optimized visual field analysis
    Tao, Yudong
    Khodeiry, Mohamed
    Ma, Rui
    Alawa, Karam
    Mendoza, Ximena
    Liu, Xiangxiang
    Shyu, Mei-Ling
    Lee, Richard K.
    INVESTIGATIVE OPHTHALMOLOGY & VISUAL SCIENCE, 2021, 62 (08)
  • [37] Deep Reinforcement Learning With Visual Attention for Vehicle Classification
    Zhao, Dongbin
    Chen, Yaran
    Lv, Le
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2017, 9 (04) : 356 - 367
  • [38] Robust reinforcement learning
    Morimoto, J
    Doya, K
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 13, 2001, 13 : 1061 - 1067
  • [39] Robust reinforcement learning
    Morimoto, J
    Doya, K
    NEURAL COMPUTATION, 2005, 17 (02) : 335 - 359
  • [40] Discriminative sampling via deep reinforcement learning for kinship verification
    Wang, Shiwei
    Yan, Haibin
    PATTERN RECOGNITION LETTERS, 2020, 138 : 38 - 43