Symbolic Task Inference in Deep Reinforcement Learning

被引:0
|
作者
Hasanbeig, Hosein [1 ]
Jeppu, Natasha Yogananda [2 ]
Abate, Alessandro [2 ]
Melham, Tom [2 ]
Kroening, Daniel [3 ]
机构
[1] Microsoft Research, United States
[2] Department of Computer Science, University of Oxford, United Kingdom
[3] Amazon, United States
关键词
Deep learning - Intelligent agents;
D O I
10.1613/jair.1.14063
中图分类号
学科分类号
摘要
This paper proposes DeepSynth, a method for effective training of deep reinforcement learning agents when the reward is sparse or non-Markovian, but at the same time progress towards the reward requires achieving an unknown sequence of high-level objectives. Our method employs a novel algorithm for synthesis of compact finite state automata to uncover this sequential structure automatically. We synthesise a human-interpretable automaton from trace data collected by exploring the environment. The state space of the environment is then enriched with the synthesised automaton, so that the generation of a control policy by deep reinforcement learning is guided by the discovered structure encoded in the automaton. The proposed approach is able to cope with both high-dimensional, low-level features and unknown sparse or non-Markovian rewards. We have evaluated DeepSynth’s performance in a set of experiments that includes the Atari game Montezuma’s Revenge, known to be challenging. Compared to approaches that rely solely on deep reinforcement learning, we obtain a reduction of two orders of magnitude in the iterations required for policy synthesis, and a significant improvement in scalability. ©2024 The Authors.
引用
收藏
页码:1099 / 1137
相关论文
共 50 条
  • [1] Symbolic Task Inference in Deep Reinforcement Learning
    Hasanbeig, Hosein
    Jeppu, Natasha Yogananda
    Abate, Alessandro
    Melham, Tom
    Kroening, Daniel
    JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2024, 80 : 1099 - 1137
  • [2] Active Task-Inference-Guided Deep Inverse Reinforcement Learning
    Memarian, Farzan
    Xu, Zhe
    Wu, Bo
    Wen, Min
    Topcu, Ufuk
    2020 59TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2020, : 1932 - 1938
  • [3] Discovering symbolic policies with deep reinforcement learning
    Landajuela, Mikel
    Petersen, Brenden K.
    Kim, Sookyung
    Santiago, Claudio P.
    Glatt, Ruben
    Mundhenk, T. Nathan
    Pettit, Jacob F.
    Faissol, Daniel M.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [4] Deep reinforcement learning with significant multiplications inference
    Ivanov, Dmitry A.
    Larionov, Denis A.
    Kiselev, Mikhail V.
    Dylov, Dmitry V.
    SCIENTIFIC REPORTS, 2023, 13 (01)
  • [5] Deep reinforcement learning with significant multiplications inference
    Dmitry A. Ivanov
    Denis A. Larionov
    Mikhail V. Kiselev
    Dmitry V. Dylov
    Scientific Reports, 13
  • [6] Deep Reinforcement Learning Task for Portfolio Construction
    Belyakov, Boris
    Sizykh, Dmitry
    21ST IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS ICDMW 2021, 2021, : 1077 - 1082
  • [7] Reinforcement Symbolic Learning
    Mercier, Chloe
    Alexandre, Frederic
    Vieville, Thierry
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2021, PT IV, 2021, 12894 : 608 - 612
  • [8] A Survey of Multi-Task Deep Reinforcement Learning
    Vithayathil Varghese, Nelson
    Mahmoud, Qusay H.
    ELECTRONICS, 2020, 9 (09) : 1 - 21
  • [9] Multi-Task Deep Reinforcement Learning with PopArt
    Hessel, Matteo
    Soyer, Hubert
    Espeholt, Lasse
    Czarnecki, Wojciech
    Schmitt, Simon
    van Hasselt, Hado
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 3796 - 3803
  • [10] Deep Reinforcement Learning for Task Offloading in Edge Computing
    Xie, Bo
    Cui, Haixia
    2024 4TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND INTELLIGENT SYSTEMS ENGINEERING, MLISE 2024, 2024, : 250 - 254