Smaller World Models for Reinforcement Learning

被引:0
|
作者
Jan Robine
Tobias Uelwer
Stefan Harmeling
机构
[1] Technical University of Dortmund,Department of Computer Science
来源
Neural Processing Letters | 2023年 / 55卷
关键词
Model-based reinforcement learning; World models; Discrete latent space; VQ-VAE; Atari;
D O I
暂无
中图分类号
学科分类号
摘要
Model-based reinforcement learning algorithms try to learn an agent by training a model that simulates the environment. However, the size of such models tends to be quite large which could be a burden as well. In this paper, we address the question, how we could design a model with fewer parameters than previous model-based approaches while achieving the same performance in the 100 K-interactions regime. For this purpose, we create a world model that combines a vector quantized-variational autoencoder to encode observations and a convolutional long short-term memory to model the dynamics. This is connected to a model-free proximal policy optimization agent to train purely on simulated experience from this world model. Detailed experiments on the Atari environments show that it is possible to reach comparable performance to the SimPLe method with a significantly smaller world model. A series of ablation studies justify our design choices and give additional insights.
引用
收藏
页码:11397 / 11427
页数:30
相关论文
共 50 条
  • [41] Reinforcement learning models for scheduling in wireless networks
    Yau, Kok-Lim Alvin
    Kwong, Kae Hsiang
    Shen, Chong
    FRONTIERS OF COMPUTER SCIENCE, 2013, 7 (05) : 754 - 766
  • [42] Models for Autonomously Motivated Exploration in Reinforcement Learning
    Auer, Peter
    Lim, Shiau Hong
    Watkins, Chris
    DISCOVERY SCIENCE, 2011, 6926 : 29 - 29
  • [43] Reinforcement Learning Models and Algorithms for Diabetes Management
    Yau, Kok-Lim Alvin
    Chong, Yung-Wey
    Fan, Xiumei
    Wu, Celimuge
    Saleem, Yasir
    Lim, Phei-Ching
    IEEE ACCESS, 2023, 11 : 28391 - 28415
  • [44] Simple reinforcement learning models are not always appropriate
    Sohn, Hansem
    Kim, Seungyeon
    JOURNAL OF NEUROSCIENCE, 2006, 26 (45): : 11511 - 11512
  • [45] The world is a smaller place
    RomanowGarcia, S
    HYDROCARBON PROCESSING, 1997, 76 (05): : 13 - 13
  • [46] Reinforcement Learning Models for Abstractive Text Summarization
    Buciumas, Sergiu
    PROCEEDINGS OF THE 2019 ANNUAL ACM SOUTHEAST CONFERENCE (ACMSE 2019), 2019, : 270 - 271
  • [47] Transferring Models in Hybrid Reinforcement Learning Agents
    Fachantidis, Anestis
    Partalas, Ioannis
    Tsoumakas, Grigorios
    Vlahavas, Ioannis
    ENGINEERING APPLICATIONS OF NEURAL NETWORKS, PT I, 2011, 363 : 162 - 171
  • [48] Fast Reinforcement Learning using Multiple Models
    Narendra, Kumpati S.
    Wang, Yu
    Mukhopadhay, Snehasis
    2016 IEEE 55TH CONFERENCE ON DECISION AND CONTROL (CDC), 2016, : 7183 - 7188
  • [49] Reinforcement learning models of aversive learning and their translation to anxiety disorders
    Seymour, Ben
    Norbury, Agnes
    JOURNAL OF NEURAL TRANSMISSION, 2017, 124 (10) : 1283 - 1284
  • [50] Integrating Neural Pathways for Learning in Deep Reinforcement Learning Models
    Ananth, Varun
    THIRTY-EIGTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 21, 2024, : 23724 - 23725