Distributionally Robust Model-based Reinforcement Learning with Large State Spaces

被引:0
|
作者
Ramesh, Shyam Sundhar [1 ]
Sessa, Pier Giuseppe [2 ]
Hu, Yifan [3 ]
Krause, Andreas [2 ]
Bogunovic, Ilija [1 ]
机构
[1] UCL, London, England
[2] Swiss Fed Inst Technol, Zurich, Switzerland
[3] Ecole Polytech Fed Lausanne, Lausanne, Switzerland
基金
英国工程与自然科学研究理事会;
关键词
MARKOV DECISION-PROCESSES;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Three major challenges in reinforcement learning are the complex dynamical systems with large state spaces, the costly data acquisition processes, and the deviation of real-world dynamics from the training environment deployment. To overcome these issues, we study distributionally robust Markov decision processes with continuous state spaces under the widely used Kullback-Leibler, chi-square, and total variation uncertainty sets. We propose a model-based approach that utilizes Gaussian Processes and the maximum variance reduction algorithm to efficiently learn multi-output nominal transition dynamics, leveraging access to a generative model (i.e., simulator). We further demonstrate the statistical sample complexity of the proposed method for different uncertainty sets. These complexity bounds are independent of the number of states and extend beyond linear dynamics, ensuring the effectiveness of our approach in identifying near-optimal distributionally-robust policies. The proposed method can be further combined with other model-free distributionally robust reinforcement learning methods to obtain a near-optimal robust policy. Experimental results demonstrate the robustness of our algorithm to distributional shifts and its superior performance in terms of the number of samples needed.
引用
收藏
页数:42
相关论文
共 50 条
  • [1] Distributionally Robust Model-Based Offline Reinforcement Learning with Near-Optimal Sample Complexity
    Shi, Laixi
    Chi, Yuejie
    JOURNAL OF MACHINE LEARNING RESEARCH, 2024, 25
  • [2] Sample Complexity of Model-Based Robust Reinforcement Learning
    Panaganti, Kishan
    Kalathil, Dileep
    2021 60TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2021, : 2240 - 2245
  • [3] A Model-Based Reinforcement Learning Approach for Robust PID Tuning
    Jesawada, Hozefa
    Yerudkar, Amol
    Del Vecchio, Carmen
    Singh, Navdeep
    2022 IEEE 61ST CONFERENCE ON DECISION AND CONTROL (CDC), 2022, : 1466 - 1471
  • [4] Robust and Explorative Behavior in Model-based Bayesian Reinforcement Learning
    Hishinuma, Toru
    Senda, Kei
    PROCEEDINGS OF 2016 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI), 2016,
  • [5] A model-based reinforcement learning approach for maintenance optimization of degrading systems in a large state space
    Zhang, Ping
    Zhu, Xiaoyan
    Xie, Min
    COMPUTERS & INDUSTRIAL ENGINEERING, 2021, 161
  • [6] Robust Model-Based Reinforcement Learning Control of a Batch Crystallization Process
    Benyahia, B.
    Anandan, P. D.
    Rielly, C.
    2021 9TH INTERNATIONAL CONFERENCE ON SYSTEMS AND CONTROL (ICSC'21), 2021, : 89 - 94
  • [7] Model-based reinforcement learning in factored-state MDPs
    Strehl, Alexander L.
    2007 IEEE INTERNATIONAL SYMPOSIUM ON APPROXIMATE DYNAMIC PROGRAMMING AND REINFORCEMENT LEARNING, 2007, : 103 - 110
  • [8] Abstract State Transition Graphs for Model-Based Reinforcement Learning
    Mendonca, Matheus R. F.
    Ziviani, Artur
    Barreto, Andre M. S.
    2018 7TH BRAZILIAN CONFERENCE ON INTELLIGENT SYSTEMS (BRACIS), 2018, : 115 - 120
  • [9] Certifiably Robust Reinforcement Learning through Model-Based Abstract Interpretation
    Yang, Chenxi
    Anderson, Greg
    Chaudhuri, Swarat
    IEEE CONFERENCE ON SAFE AND TRUSTWORTHY MACHINE LEARNING, SATML 2024, 2024, : 233 - 251
  • [10] Risk-Averse Model Uncertainty for Distributionally Robust Safe Reinforcement Learning
    Queeney, James
    Benosman, Mouhacine
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,