Planning with Abstract Markov Decision Processes

被引:0
|
作者
Gopalan, Nakul [1 ]
desJardins, Marie [2 ]
Littman, Michael L. [1 ]
MacGlashan, James [3 ]
Squire, Shawn [2 ]
Tellex, Stefanie [1 ]
Winder, John [2 ]
Wong, Lawson L. S. [1 ]
机构
[1] Brown Univ, Providence, RI 02912 USA
[2] Univ Maryland Baltimore Cty, Baltimore, MD 21250 USA
[3] Cogitai Inc, Riverside, RI USA
基金
美国国家科学基金会; 美国国家航空航天局;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Robots acting in human-scale environments must plan under uncertainty in large state-action spaces and face constantly changing reward functions as requirements and goals change. Planning under uncertainty in large state-action spaces requires hierarchical abstraction for efficient computation. We introduce a new hierarchical planning framework called Abstract Markov Decision Processes (AMDPs) that can plan in a fraction of the time needed for complex decision making in ordinary MDPs. AMDPs provide abstract states, actions, and transition dynamics in multiple layers above a base-level "flat" MDP. AMDPs decompose problems into a series of subtasks with both local reward and local transition functions used to create policies for subtasks. The resulting hierarchical planning method is independently optimal at each level of abstraction, and is recursively optimal when the local reward and transition functions are correct. We present empirical results showing significantly improved planning speed, while maintaining solution quality, in the Taxi domain and in a mobile-manipulation robotics problem. Furthermore, our approach allows specification of a decision-making model for a mobile-manipulation problem on a Turtlebot, spanning from low-level control actions operating on continuous variables all the way up through high-level object manipulation tasks.
引用
收藏
页码:480 / 488
页数:9
相关论文
共 50 条
  • [31] Inspection and maintenance planning: an application of semi-Markov decision processes
    Universite de Technologie de Troyes, Troyes, France
    J Intell Manuf, 5 (467-476):
  • [33] Minimax Regret Optimisation for Robust Planning in Uncertain Markov Decision Processes
    Rigter, Marc
    Lacerda, Bruno
    Hawes, Nick
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 11930 - 11938
  • [34] Strategic Planning under Uncertainties via Constrained Markov Decision Processes
    Ding, Xu Chu
    Pinto, Alessandro
    Surana, Amit
    2013 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2013, : 4568 - 4575
  • [35] Inspection and maintenance planning: an application of semi-Markov decision processes
    Berenguer, C
    Chu, CB
    Grall, A
    JOURNAL OF INTELLIGENT MANUFACTURING, 1997, 8 (05) : 467 - 476
  • [36] Inspection and maintenance planning: an application of semi-Markov decision processes
    CHRISTOPHE BERENGUER
    CHENGBIN CHU
    ANTOINE GRALL
    Journal of Intelligent Manufacturing, 1997, 8 : 467 - 476
  • [37] Policy Reuse for Learning and Planning in Partially Observable Markov Decision Processes
    Wu, Bo
    Feng, Yanpeng
    2017 4TH INTERNATIONAL CONFERENCE ON INFORMATION SCIENCE AND CONTROL ENGINEERING (ICISCE), 2017, : 549 - 552
  • [38] Global path planning for AUV based on hierarchical Markov decision processes
    Hong, Ye
    Wang, Hong-Jian
    Bian, Xin-Qian
    Xitong Fangzhen Xuebao / Journal of System Simulation, 2008, 20 (09): : 2361 - 2363
  • [39] Markov Decision Processes For Multi-Objective Satellite Task Planning
    Eddy, Duncan
    Kochenderfer, Mykel
    2020 IEEE AEROSPACE CONFERENCE (AEROCONF 2020), 2020,
  • [40] Learning Sequential Decision Tasks for Robot Manipulation with Abstract Markov Decision Processes and Demonstration-Guided Exploration
    Kent, David
    Banerjee, Siddhartha
    Chernova, Sonia
    2018 IEEE-RAS 18TH INTERNATIONAL CONFERENCE ON HUMANOID ROBOTS (HUMANOIDS), 2018, : 958 - 965