PROBING TRANSFER IN DEEP REINFORCEMENT LEARNING WITHOUT TASK ENGINEERING

被引:0
|
作者
Rusu, Andrei A. [1 ]
Flennerhag, Sebastian [1 ]
Rao, Dushyant [1 ]
Pascanu, Razvan [1 ]
Hadsell, Raia [1 ]
机构
[1] DeepMind, London, England
来源
CONFERENCE ON LIFELONG LEARNING AGENTS, VOL 199 | 2022年 / 199卷
关键词
GO; ENVIRONMENT; LEVEL; GAME;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We evaluate the use of original game curricula supported by the Atari 2600 console as a heterogeneous transfer benchmark for deep reinforcement learning agents. Game designers created curricula using combinations of several discrete modifications to the basic versions of games such as Space Invaders, Breakout and Freeway, making them progressively more challenging for human players. By formally organising these modifications into several factors of variation, we are able to show that Analyses of Variance (ANOVA) are a potent tool for studying the effects of human-relevant domain changes on the learning and transfer performance of a deep reinforcement learning agent. Since no manual task engineering is needed on our part, leveraging the original multi-factorial design avoids the pitfalls of unintentionally biasing the experimental setup. We find that game design factors have a large and statistically significant impact on an agent's ability to learn, and so do their combinatorial interactions. Furthermore, we show that zero-shot transfer from the basic games to their respective variations is possible, but the variance in performance is also largely explained by interactions between factors. As such, we argue that Atari game curricula offer a challenging benchmark for transfer learning in RL, that can help the community better understand the generalisation capabilities of RL agents along dimensions which meaningfully impact human generalisation performance. As a start, we report that value-function finetuning of regularly trained agents achieves positive transfer in a majority of cases, but significant headroom for algorithmic innovation remains. We conclude with the observation that selective transfer from multiple variants could further improve performance.
引用
收藏
页数:24
相关论文
共 50 条
  • [41] Multi-task Deep Reinforcement Learning for Scalable Parallel Task Scheduling
    Zhang, Lingxin
    Qi, Qi
    Wang, Jingyu
    Sun, Haifeng
    Liao, Jianxin
    2019 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2019, : 2992 - 3001
  • [42] The Effect of Task Similarity on Deep Transfer Learning
    Zhang, Wei
    Fang, Yuchun
    Ma, Zhengyan
    NEURAL INFORMATION PROCESSING (ICONIP 2017), PT II, 2017, 10635 : 256 - 265
  • [43] Distributed and Adaptive Traffic Engineering with Deep Reinforcement Learning
    Geng, Nan
    Xu, Mingwei
    Yang, Yuan
    Liu, Chenyi
    Yang, Jiahai
    Li, Qi
    Zhang, Shize
    2021 IEEE/ACM 29TH INTERNATIONAL SYMPOSIUM ON QUALITY OF SERVICE (IWQOS), 2021,
  • [44] Feature Engineering for Deep Reinforcement Learning Based Routing
    Suarez-Varela, Jose
    Mestres, Albert
    Yu, Junlin
    Kuang, Li
    Feng, Haoyu
    Barlet-Ros, Pere
    Cabellos-Aparicio, Albert
    ICC 2019 - 2019 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC), 2019,
  • [45] Leveraging Deep Reinforcement Learning for Traffic Engineering: A Survey
    Xiao, Yang
    Liu, Jun
    Wu, Jiawei
    Ansari, Nirwan
    IEEE COMMUNICATIONS SURVEYS AND TUTORIALS, 2021, 23 (04): : 2064 - 2097
  • [46] Multi-source Transfer Learning for Deep Reinforcement Learning
    Garcia-Ramirez, Jesus
    Morales, Eduardo
    Escalante, Hugo Jair
    PATTERN RECOGNITION (MCPR 2021), 2021, 12725 : 131 - 140
  • [47] Spectrum-Agile Cognitive Radios Using Multi-Task Transfer Deep Reinforcement Learning
    Aref, Mohamed A.
    Jayaweera, Sudharman K.
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2021, 20 (10) : 6729 - 6742
  • [48] Deep Reinforcement Learning Task Assignment Based on Domain Knowledge
    Liu, Jiayi
    Wang, Gang
    Guo, Xiangke
    Wang, Siyuan
    Fu, Qiang
    IEEE Access, 2022, 10 : 114402 - 114413
  • [49] Accelerating Deep Continuous Reinforcement Learning through Task Simplification
    Kerzel, Matthias
    Mohammadi, Hadi Beik
    Zamani, Mohammad Ali
    Wermter, Stefan
    2018 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2018, : 139 - 144
  • [50] Task Assignment of UAV Swarms Based on Deep Reinforcement Learning
    Liu, Bo
    Wang, Shulei
    Li, Qinghua
    Zhao, Xinyang
    Pan, Yunqing
    Wang, Changhong
    DRONES, 2023, 7 (05)