Variational Inference MPC for Bayesian Model-based Reinforcement Learning

被引:0
|
作者
Okada, Masashi [1 ]
Taniguchi, Tadahiro [1 ,2 ]
机构
[1] Panason Corp, Kadoma, Osaka, Japan
[2] Ritsumeikan Univ, Kyoto, Japan
来源
关键词
model predictive control; variational inference; model-based reinforcement learning; PREDICTIVE CONTROL; OPTIMIZATION;
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
In recent studies on model-based reinforcement learning (MBRL), incorporating uncertainty in forward dynamics is a state-of-the-art strategy to enhance learning performance, making MBRLs competitive to cutting-edge model-free methods, especially in simulated robotics tasks. Probabilistic ensembles with trajectory sampling (PETS) is a leading type of MBRL, which employs Bayesian inference to dynamics modeling and model predictive control (MPC) with stochastic optimization via the cross entropy method (CEM). In this paper, we propose a novel extension to the uncertainty-aware MBRL. Our main contributions are twofold: Firstly, we introduce a variational inference MPC (VI-MPC), which reformulates various stochastic methods, including CEM, in a Bayesian fashion. Secondly, we propose a novel instance of the framework, called probabilistic action ensembles with trajectory sampling (PaETS). As a result, our Bayesian MBRL can involve multimodal uncertainties both in dynamics and optimal trajectories. In comparison to PETS, our method consistently improves asymptotic performance on several challenging locomotion tasks.
引用
收藏
页数:15
相关论文
共 50 条
  • [31] VIREL: A Variational Inference Framework for Reinforcement Learning
    Fellows, Matthew
    Mahajan, Anuj
    Rudner, Tim G. J.
    Whiteson, Shimon
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [32] Learning to Paint With Model-based Deep Reinforcement Learning
    Huang, Zhewei
    Heng, Wen
    Zhou, Shuchang
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 8708 - 8717
  • [33] MRF model-based joint interrupted SAR imaging and coherent change detection via variational Bayesian inference
    Yang, Yue
    Cong, Xunchao
    Long, Keyu
    Luo, Yongjie
    Xie, Wei
    Wan, Qun
    SIGNAL PROCESSING, 2018, 151 : 144 - 154
  • [34] Variational Bayesian Reinforcement Learning with Regret Bounds
    O'Donoghue, Brendan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [35] Incremental model-based reinforcement learning with model constraint
    Yang, Zhiyou
    Fu, Mingsheng
    Qu, Hong
    Li, Fan
    Shi, Shuqing
    Hu, Wang
    NEURAL NETWORKS, 2025, 185
  • [36] Model-based Policy Optimization under Approximate Bayesian Inference
    Wang, Chaoqi
    Chen, Yuxin
    Murphy, Kevin
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 238, 2024, 238
  • [37] Bayesian inference for model-based segmentation of computed radiographs of the hand
    Levitt, T.S.
    Hedgcock Jr., M.W.
    Dye, J.W.
    Johnston, S.E.
    Shadle, V.M.
    Vosky, D.
    Artificial Intelligence in Medicine, 1993, 5 (04) : 365 - 387
  • [38] A Sparse Bayesian Learning Method for Structural Equation Model-Based Gene Regulatory Network Inference
    Li, Yan
    Liu, Dayou
    Chu, Jianfeng
    Zhu, Yungang
    Liu, Jie
    Cheng, Xiaochun
    IEEE ACCESS, 2020, 8 : 40067 - 40080
  • [39] Online reinforcement learning control by Bayesian inference
    Xia, Zhongpu
    Zhao, Dongbin
    IET CONTROL THEORY AND APPLICATIONS, 2016, 10 (12): : 1331 - 1338
  • [40] Combining deep learning and model-based method using Bayesian Inference for walking speed estimation
    Qian, Yuyang
    Yang, Kaiming
    Zhu, Yu
    Wang, Wei
    Wan, Chenhui
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2020, 62