Planning in Discrete and Continuous Markov Decision Processes by Probabilistic Programming

被引:7
|
作者
Nitti, Davide [1 ]
Belle, Vaishak [1 ]
de Raedt, Luc [1 ]
机构
[1] KU, Dept Comp Sci, Leuven, Belgium
关键词
D O I
10.1007/978-3-319-23525-7_20
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Real-world planning problems frequently involve mixtures of continuous and discrete state variables and actions, and are formulated in environments with an unknown number of objects. In recent years, probabilistic programming has emerged as a natural approach to capture and characterize such complex probability distributions with general-purpose inference methods. While it is known that a probabilistic programming language can be easily extended to represent Markov Decision Processes (MDPs) for planning tasks, solving such tasks is challenging. Building on related efforts in reinforcement learning, we introduce a conceptually simple but powerful planning algorithm for MDPs realized as a probabilistic program. This planner constructs approximations to the optimal policy by importance sampling, while exploiting the knowledge of the MDP model. In our empirical evaluations, we show that this approach has wide applicability on domains ranging from strictly discrete to strictly continuous to hybrid ones, handles intricacies such as unknown objects, and is argued to be competitive given its generality.
引用
收藏
页码:327 / 342
页数:16
相关论文
共 50 条
  • [1] Probabilistic Preference Planning Problem for Markov Decision Processes
    Li, Meilun
    Turrini, Andrea
    Hahn, Ernst Moritz
    She, Zhikun
    Zhang, Lijun
    IEEE TRANSACTIONS ON SOFTWARE ENGINEERING, 2022, 48 (05) : 1545 - 1559
  • [2] Markov Decision Processes Specified by Probabilistic Logic Programming: Representation and Solution
    Bueno, Thiago P.
    Maua, Denis D.
    de Barros, Leliane N.
    Cozman, Fabio G.
    PROCEEDINGS OF 2016 5TH BRAZILIAN CONFERENCE ON INTELLIGENT SYSTEMS (BRACIS 2016), 2016, : 337 - 342
  • [3] Optimal control of probabilistic discrete event systems on Markov decision processes
    Wang, Fei
    Feng, Zu-Ren
    Hu, Qi-Ying
    Kongzhi Lilun Yu Yingyong/Control Theory and Applications, 2007, 24 (06): : 895 - 901
  • [4] Unifying nondeterministic and probabilistic planning through imprecise Markov Decision Processes
    Trevizan, Felipe W.
    Cozman, Fabio G.
    de Barros, Leliane N.
    ADVANCES IN ARTIFICIAL INTELLIGENCE - IBERAMIA-SBIA 2006, PROCEEDINGS, 2006, 4140 : 502 - 511
  • [5] Mean Field for Markov Decision Processes: From Discrete to Continuous Optimization
    Gast, Nicolas
    Gaujal, Bruno
    Le Boudec, Jean-Yves
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2012, 57 (09) : 2266 - 2280
  • [6] Probabilistic Hyperproperties of Markov Decision Processes
    Dimitrova, Rayna
    Finkbeiner, Bernd
    Torfah, Hazem
    AUTOMATED TECHNOLOGY FOR VERIFICATION AND ANALYSIS (ATVA 2020), 2020, 12302 : 484 - 500
  • [7] Probabilistic opacity for Markov decision processes
    Berard, Beatrice
    Chatterjee, Krishnendu
    Sznajder, Nathalie
    INFORMATION PROCESSING LETTERS, 2015, 115 (01) : 52 - 59
  • [8] Quantitative Programming and Markov Decision Processes
    Todoran, Eneia Nicolae
    2022 24TH INTERNATIONAL SYMPOSIUM ON SYMBOLIC AND NUMERIC ALGORITHMS FOR SCIENTIFIC COMPUTING, SYNASC, 2022, : 117 - 124
  • [9] Markov reward models and markov decision processes in discrete and continuous time: Performance evaluation and optimization
    Gouberman, Alexander
    Siegle, Markus
    Gouberman, Alexander (alexander.gouberman@unibw.de), 1600, Springer Verlag (8453): : 156 - 241
  • [10] EQUIVALENCE BETWEEN CONTINUOUS AND DISCRETE-TIME MARKOV DECISION-PROCESSES
    SERFOZO, RF
    OPERATIONS RESEARCH, 1979, 27 (03) : 616 - 620