Variational Inference MPC for Bayesian Model-based Reinforcement Learning

被引:0
|
作者
Okada, Masashi [1 ]
Taniguchi, Tadahiro [1 ,2 ]
机构
[1] Panason Corp, Kadoma, Osaka, Japan
[2] Ritsumeikan Univ, Kyoto, Japan
来源
关键词
model predictive control; variational inference; model-based reinforcement learning; PREDICTIVE CONTROL; OPTIMIZATION;
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
In recent studies on model-based reinforcement learning (MBRL), incorporating uncertainty in forward dynamics is a state-of-the-art strategy to enhance learning performance, making MBRLs competitive to cutting-edge model-free methods, especially in simulated robotics tasks. Probabilistic ensembles with trajectory sampling (PETS) is a leading type of MBRL, which employs Bayesian inference to dynamics modeling and model predictive control (MPC) with stochastic optimization via the cross entropy method (CEM). In this paper, we propose a novel extension to the uncertainty-aware MBRL. Our main contributions are twofold: Firstly, we introduce a variational inference MPC (VI-MPC), which reformulates various stochastic methods, including CEM, in a Bayesian fashion. Secondly, we propose a novel instance of the framework, called probabilistic action ensembles with trajectory sampling (PaETS). As a result, our Bayesian MBRL can involve multimodal uncertainties both in dynamics and optimal trajectories. In comparison to PETS, our method consistently improves asymptotic performance on several challenging locomotion tasks.
引用
收藏
页数:15
相关论文
共 50 条
  • [1] Importance-Weighted Variational Inference Model Estimation for Offline Bayesian Model-Based Reinforcement Learning
    Hishinuma, Toru
    Senda, Kei
    IEEE ACCESS, 2023, 11 : 145579 - 145590
  • [2] Direct policy search reinforcement learning based on variational Bayesian inference
    Yamaguchi, Nobuhiko
    Ihara, Kazuya
    Fukuda, Osamu
    Okumura, Hiroshi
    2018 JOINT 10TH INTERNATIONAL CONFERENCE ON SOFT COMPUTING AND INTELLIGENT SYSTEMS (SCIS) AND 19TH INTERNATIONAL SYMPOSIUM ON ADVANCED INTELLIGENT SYSTEMS (ISIS), 2018, : 1009 - 1014
  • [3] Direct Policy Search Reinforcement Learning Based on Variational Bayesian Inference
    Yamaguchi, Nobuhiko
    JOURNAL OF ADVANCED COMPUTATIONAL INTELLIGENCE AND INTELLIGENT INFORMATICS, 2020, 24 (06) : 711 - 718
  • [4] Model-based Bayesian Reinforcement Learning for Dialogue Management
    Lison, Pierre
    14TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2013), VOLS 1-5, 2013, : 475 - 479
  • [5] Model-based Lifelong Reinforcement Learning with Bayesian Exploration
    Fu, Haotian
    Yu, Shangqun
    Littman, Michael
    Konidaris, George
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [6] Smarter Sampling in Model-Based Bayesian Reinforcement Learning
    Castro, Pablo Samuel
    Precup, Doina
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, PT I: EUROPEAN CONFERENCE, ECML PKDD 2010, 2010, 6321 : 200 - 214
  • [7] A Model-based Factored Bayesian Reinforcement Learning Approach
    Wu, Bo
    Feng, Yanpeng
    Zheng, Hongyan
    APPLIED SCIENCE, MATERIALS SCIENCE AND INFORMATION TECHNOLOGIES IN INDUSTRY, 2014, 513-517 : 1092 - 1095
  • [8] Reward Shaping for Model-Based Bayesian Reinforcement Learning
    Kim, Hyeoneun
    Lim, Woosang
    Lee, Kanghoon
    Noh, Yung-Kyun
    Kim, Kee-Eung
    PROCEEDINGS OF THE TWENTY-NINTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2015, : 3548 - 3555
  • [9] Learning safety in model-based Reinforcement Learning using MPC and Gaussian Processes
    Airaldi, Filippo
    De Schutter, Bart
    Dabiri, Azita
    IFAC PAPERSONLINE, 2023, 56 (02): : 5759 - 5764
  • [10] Kernel-based direct policy search reinforcement learning based on variational Bayesian inference
    Yamaguchi, Nobuhiko
    Fukuda, Osamu
    Okumura, Hiroshi
    2019 SEVENTH INTERNATIONAL SYMPOSIUM ON COMPUTING AND NETWORKING WORKSHOPS (CANDARW 2019), 2019, : 184 - 187