Oracle-SAGE: Planning Ahead in Graph-Based Deep Reinforcement Learning

被引:1
|
作者
Chester, Andrew [1 ]
Dann, Michael [1 ]
Zambetta, Fabio [1 ]
Thangarajah, John [1 ]
机构
[1] RMIT Univ, Sch Comp Technol, Melbourne, Australia
关键词
Reinforcement learning; GNNs; Symbolic planning;
D O I
10.1007/978-3-031-26412-2_4
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep reinforcement learning (RL) commonly suffers from high sample complexity and poor generalisation, especially with high-dimensional (image-based) input. Where available (such as some robotic control domains), low dimensional vector inputs outperform their image based counterparts, but it is challenging to represent complex dynamic environments in this manner. Relational reinforcement learning instead represents the world as a set of objects and the relations between them; offering a flexible yet expressive view which provides structural inductive biases to aid learning. Recently relational RL methods have been extended with modern function approximation using graph neural networks (GNNs). However, inherent limitations in the processing model for GNNs result in decreased returns when important information is dispersed widely throughout the graph. We outline a hybrid learning and planning model which uses reinforcement learning to propose and select subgoals for a planning model to achieve. This includes a novel action selection mechanism and loss function to allow training around the non-differentiable planner. We demonstrate our algorithms effectiveness on a range of domains, including MiniHack and a challenging extension of the classic taxi domain.
引用
收藏
页码:52 / 67
页数:16
相关论文
共 50 条
  • [41] Graph-based rank aggregation: a deep-learning approach
    Keyhanipour, Amir Hosein
    INTERNATIONAL JOURNAL OF WEB INFORMATION SYSTEMS, 2025, 21 (01) : 54 - 76
  • [42] kGCN: a graph-based deep learning framework for chemical structures
    Kojima, Ryosuke
    Ishida, Shoichi
    Ohta, Masateru
    Iwata, Hiroaki
    Honma, Teruki
    Okuno, Yasushi
    JOURNAL OF CHEMINFORMATICS, 2020, 12 (01)
  • [43] kGCN: a graph-based deep learning framework for chemical structures
    Ryosuke Kojima
    Shoichi Ishida
    Masateru Ohta
    Hiroaki Iwata
    Teruki Honma
    Yasushi Okuno
    Journal of Cheminformatics, 12
  • [44] De novo drug design by iterative multiobjective deep reinforcement learning with graph-based molecular quality assessment
    Fang, Yi
    Pan, Xiaoyong
    Shen, Hong-Bin
    BIOINFORMATICS, 2023, 39 (04)
  • [45] Collaborative multi-target-tracking via graph-based deep reinforcement learning in UAV swarm networks
    Ren, Qianchen
    Wang, Yuanyu
    Liu, Han
    Dai, Yu
    Ye, Wenhui
    Tang, Yuliang
    AD HOC NETWORKS, 2025, 172
  • [46] Learning Heterogeneous Strategies via Graph-based Multi-agent Reinforcement Learning
    Li, Yang
    Luo, Xiangfeng
    Xie, Shaorong
    2021 IEEE 33RD INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2021), 2021, : 709 - 713
  • [47] Energy Efficient UAV-Assisted IoT Data Collection: A Graph-Based Deep Reinforcement Learning Approach
    Wu, Qianqian
    Liu, Qiang
    Zhu, Wenliang
    Wu, Zefan
    IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT, 2024, 21 (06): : 6082 - 6094
  • [48] Enhancing Federated Learning Performance Fairness via Collaboration Graph-Based Reinforcement Learning
    Xia, Yuexuan
    Ma, Benteng
    Dou, Qi
    Xia, Yong
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2024, PT X, 2024, 15010 : 263 - 272
  • [49] Assembly sequence planning based on deep reinforcement learning
    Zhao M.-H.
    Zhang X.-B.
    Guo X.
    Ou Y.-S.
    Kongzhi Lilun Yu Yingyong/Control Theory and Applications, 2021, 38 (12): : 1901 - 1910
  • [50] Robot path planning based on deep reinforcement learning
    Long, Yinxin
    He, Huajin
    2020 IEEE CONFERENCE ON TELECOMMUNICATIONS, OPTICS AND COMPUTER SCIENCE (TOCS), 2020, : 151 - 154