Symbolic Task Inference in Deep Reinforcement Learning

被引:0
|
作者
Hasanbeig, Hosein [1 ]
Jeppu, Natasha Yogananda [2 ]
Abate, Alessandro [2 ]
Melham, Tom [2 ]
Kroening, Daniel [3 ]
机构
[1] Microsoft Research, United States
[2] Department of Computer Science, University of Oxford, United Kingdom
[3] Amazon, United States
关键词
Deep learning - Intelligent agents;
D O I
10.1613/jair.1.14063
中图分类号
学科分类号
摘要
This paper proposes DeepSynth, a method for effective training of deep reinforcement learning agents when the reward is sparse or non-Markovian, but at the same time progress towards the reward requires achieving an unknown sequence of high-level objectives. Our method employs a novel algorithm for synthesis of compact finite state automata to uncover this sequential structure automatically. We synthesise a human-interpretable automaton from trace data collected by exploring the environment. The state space of the environment is then enriched with the synthesised automaton, so that the generation of a control policy by deep reinforcement learning is guided by the discovered structure encoded in the automaton. The proposed approach is able to cope with both high-dimensional, low-level features and unknown sparse or non-Markovian rewards. We have evaluated DeepSynth’s performance in a set of experiments that includes the Atari game Montezuma’s Revenge, known to be challenging. Compared to approaches that rely solely on deep reinforcement learning, we obtain a reduction of two orders of magnitude in the iterations required for policy synthesis, and a significant improvement in scalability. ©2024 The Authors.
引用
收藏
页码:1099 / 1137
相关论文
共 50 条
  • [41] Task Inference for Offline Meta Reinforcement Learning via Latent Shared Knowledge
    Zhou, Ying
    Cong, Shan
    Yu, Chao
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT IV, KSEM 2023, 2023, 14120 : 356 - 365
  • [42] Adaptive Inference Reinforcement Learning for Task Offloading in Vehicular Edge Computing Systems
    Tang, Dian
    Zhang, Xuefei
    Li, Meng
    Tao, Xiaofeng
    2020 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS WORKSHOPS (ICC WORKSHOPS), 2020,
  • [43] A CGRA based Neural Network Inference Engine for Deep Reinforcement Learning
    Liang, Minglan
    Chen, Mingsong
    Wang, Zheng
    Sun, Jingwei
    2018 IEEE ASIA PACIFIC CONFERENCE ON CIRCUITS AND SYSTEMS (APCCAS 2018), 2018, : 540 - 543
  • [44] Deep reinforcement learning for task-based feature learning in prosthetic vision
    White, Jack
    Kameneva, Tatiana
    McCarthy, Chris
    2019 41ST ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY (EMBC), 2019, : 2809 - 2812
  • [45] Deep Reinforcement Learning Based Resource Management for DNN Inference in IIoT
    Zhang, Weiting
    Yang, Dong
    Peng, Haixia
    Wu, Wen
    Quan, Wei
    Zhang, Hongke
    Shen, Xuemin
    2020 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2020,
  • [46] Spatiotemporal Costmap Inference for MPC Via Deep Inverse Reinforcement Learning
    Lee, Keuntaek
    Isele, David
    Theodorou, Evangelos A.
    Bae, Sangjae
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (02) : 3194 - 3201
  • [47] Substation Operation Sequence Inference Model Based on Deep Reinforcement Learning
    Chen, Tie
    Li, Hongxin
    Cao, Ying
    Zhang, Zhifan
    APPLIED SCIENCES-BASEL, 2023, 13 (13):
  • [48] Zero-Shot Task Generalization with Multi-Task Deep Reinforcement Learning
    Oh, Junhyuk
    Singh, Satinder
    Lee, Honglak
    Kohli, Pushmeet
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
  • [49] Creativity of AI: Automatic Symbolic Option Discovery for Facilitating Deep Reinforcement Learning
    Jin, Mu
    Ma, Zhihao
    Jin, Kebing
    Zhuo, Hankz Hankui
    Chen, Chen
    Yu, Chao
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 7042 - 7050
  • [50] Symbolic Task Compression in Structured Task Learning
    Saveriano, Matteo
    Seegerer, Michael
    Caccavale, Riccardo
    Finzi, Alberto
    Lee, Dongheui
    2019 THIRD IEEE INTERNATIONAL CONFERENCE ON ROBOTIC COMPUTING (IRC 2019), 2019, : 171 - 176