GRIMGEP: Learning Progress for Robust Goal Sampling in Visual Deep Reinforcement Learning

被引:2
|
作者
Kovac, Grgur [1 ]
Laversanne-Finot, Adrien [1 ]
Oudeyer, Pierre-Yves [1 ]
机构
[1] INRIA Bordeaux, Flowers Lab, F-33400 Talence, France
关键词
Goal exploration; learning progress; reinforcement learning (RL); INTRINSIC MOTIVATION; EXPLORATION; SYSTEMS;
D O I
10.1109/TCDS.2022.3216911
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Autotelic reinforcement learning (RL) agents sample their own goals, and try to reach them. They often prioritize goal sampling according to some intrinsic reward, ex. novelty or absolute learning progress (ALPs). Novelty-based approaches work robustly in unsupervised image-based environments when there are no distractors. However, they construct simple curricula that do not take the agent's performance into account: in complex environments, they often get attracted by impossible tasks. ALP-based approaches, which are often combined with a clustering mechanism, construct complex curricula tuned to the agent's current capabilities. Such curricula sample goals on which the agent is currently learning the most, and do not get attracted by impossible tasks. However, ALP approaches have not so far been applied to DRL agents perceiving complex environments directly in the image space. Goal regions guided intrinsically motivated goal exploration process (GRIMGEP), without using any expert knowledge, combines the ALP clustering approaches with novelty-based approaches and extends them to those complex scenarios. We experiment on a rich 3-D image-based environment with distractors using novelty-based exploration approaches: Skewfit and CountBased. We show that wrapping them with GRIMGEP-using them only in the cluster sampled by ALP-creates a better curriculum. The wrapped approaches are attracted less by the distractors, and achieve drastically better performances.
引用
收藏
页码:1396 / 1407
页数:12
相关论文
共 50 条
  • [41] Goal- Driven Autonomous Exploration Through Deep Reinforcement Learning
    Cimurs, Reinis
    Suh, Il Hong
    Lee, Jin Han
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (02) : 730 - 737
  • [42] Robust Reinforcement Learning Based Visual Servoing with Convolutional Features
    Fei, Haolin
    Wang, Ziwei
    Kennedy, Andrew
    IFAC PAPERSONLINE, 2023, 56 (02): : 9781 - 9786
  • [43] Evaluating the progress of deep learning for visual relational concepts
    Stabinger, Sebastian
    Peer, David
    Piater, Justus
    Rodriguez-Sanchez, Antonio
    JOURNAL OF VISION, 2021, 21 (11): : 11 - 23
  • [44] Contrastive Learning as Goal-Conditioned Reinforcement Learning
    Eysenbach, Benjamin
    Zhang, Tianjun
    Levine, Sergey
    Salakhutdinov, Ruslan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [45] Research progress of deep reinforcement learning applied to text generation
    Xu C.
    Li Q.
    Zhang D.-Z.
    Chen P.
    Cui J.-R.
    Gongcheng Kexue Xuebao/Chinese Journal of Engineering, 2020, 42 (04): : 399 - 411
  • [46] Research Progress of Multi-Agent Deep Reinforcement Learning
    Ding, Shi-Feiu
    Du, Weiu
    Zhang, Jianu
    Guo, Li-Liu
    Ding, Ding
    Jisuanji Xuebao/Chinese Journal of Computers, 2024, 47 (07): : 1547 - 1567
  • [47] DEEP LEARNING FOR ROBUST OUTDOOR VEHICLE VISUAL TRACKING
    Xin, Jing
    Du, Xing
    Zhang, Jian
    2017 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2017, : 613 - 618
  • [48] SparseIDS: Learning Packet Sampling with Reinforcement Learning
    Bachl, Maximilian
    Meghdouri, Fares
    Fabini, Joachim
    Zseby, Tanja
    2020 IEEE CONFERENCE ON COMMUNICATIONS AND NETWORK SECURITY (CNS), 2020,
  • [49] Achieving Robust Learning Outcomes in Autonomous Driving with DynamicNoise Integration in Deep Reinforcement Learning
    Shi, Haotian
    Chen, Jiale
    Zhang, Feijun
    Liu, Mingyang
    Zhou, Mengjie
    DRONES, 2024, 8 (09)
  • [50] RSAC: A Robust Deep Reinforcement Learning Strategy for Dimensionality Perturbation
    Gupta, Surbhi
    Singal, Gaurav
    Garg, Deepak
    Das, Swagatam
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2022, 6 (05): : 1157 - 1166