Planning with Abstract Markov Decision Processes

被引:0
|
作者
Gopalan, Nakul [1 ]
desJardins, Marie [2 ]
Littman, Michael L. [1 ]
MacGlashan, James [3 ]
Squire, Shawn [2 ]
Tellex, Stefanie [1 ]
Winder, John [2 ]
Wong, Lawson L. S. [1 ]
机构
[1] Brown Univ, Providence, RI 02912 USA
[2] Univ Maryland Baltimore Cty, Baltimore, MD 21250 USA
[3] Cogitai Inc, Riverside, RI USA
基金
美国国家科学基金会; 美国国家航空航天局;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Robots acting in human-scale environments must plan under uncertainty in large state-action spaces and face constantly changing reward functions as requirements and goals change. Planning under uncertainty in large state-action spaces requires hierarchical abstraction for efficient computation. We introduce a new hierarchical planning framework called Abstract Markov Decision Processes (AMDPs) that can plan in a fraction of the time needed for complex decision making in ordinary MDPs. AMDPs provide abstract states, actions, and transition dynamics in multiple layers above a base-level "flat" MDP. AMDPs decompose problems into a series of subtasks with both local reward and local transition functions used to create policies for subtasks. The resulting hierarchical planning method is independently optimal at each level of abstraction, and is recursively optimal when the local reward and transition functions are correct. We present empirical results showing significantly improved planning speed, while maintaining solution quality, in the Taxi domain and in a mobile-manipulation robotics problem. Furthermore, our approach allows specification of a decision-making model for a mobile-manipulation problem on a Turtlebot, spanning from low-level control actions operating on continuous variables all the way up through high-level object manipulation tasks.
引用
收藏
页码:480 / 488
页数:9
相关论文
共 50 条
  • [1] Abstract interpretation of programs as Markov decision processes
    Monniaux, D
    SCIENCE OF COMPUTER PROGRAMMING, 2005, 58 (1-2) : 179 - 205
  • [2] Abstract interpretation of programs as Markov decision processes
    Monniaux, D
    STATIC ANALYSIS, PROCEEDINGS, 2003, 2694 : 237 - 254
  • [3] Preference Planning for Markov Decision Processes
    Li, Meilun
    She, Zhikun
    Turrini, Andrea
    Zhang, Lijun
    PROCEEDINGS OF THE TWENTY-NINTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2015, : 3313 - 3319
  • [4] Multiagent, Multitarget Path Planning in Markov Decision Processes
    Nawaz, Farhad
    Ornik, Melkior
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2023, 68 (12) : 7560 - 7574
  • [5] Approximate planning and verification for large Markov decision processes
    Lassaigne, Richard
    Peyronnet, Sylvain
    INTERNATIONAL JOURNAL ON SOFTWARE TOOLS FOR TECHNOLOGY TRANSFER, 2015, 17 (04) : 457 - 467
  • [6] Oblivious Markov Decision Processes: Planning and Policy Execution
    Alsayegh, Murtadha
    Fuentes, Jose
    Bobadilla, Leonardo
    Shell, Dylan A.
    2023 62ND IEEE CONFERENCE ON DECISION AND CONTROL, CDC, 2023, : 3850 - 3857
  • [7] Planning using hierarchical constrained Markov decision processes
    Seyedshams Feyzabadi
    Stefano Carpin
    Autonomous Robots, 2017, 41 : 1589 - 1607
  • [8] Planning using hierarchical constrained Markov decision processes
    Feyzabadi, Seyedshams
    Carpin, Stefano
    AUTONOMOUS ROBOTS, 2017, 41 (08) : 1589 - 1607
  • [9] Probabilistic Preference Planning Problem for Markov Decision Processes
    Li, Meilun
    Turrini, Andrea
    Hahn, Ernst Moritz
    She, Zhikun
    Zhang, Lijun
    IEEE TRANSACTIONS ON SOFTWARE ENGINEERING, 2022, 48 (05) : 1545 - 1559
  • [10] Learning and Planning with Timing Information in Markov Decision Processes
    Bacon, Pierre-Luc
    Balle, Borja
    Precup, Doina
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2015, : 111 - 120