Monte Carlo Hierarchical Model Learning

被引:0
|
作者
Menashe, Jacob [1 ]
Stone, Peter [1 ]
机构
[1] Univ Texas Austin, Austin, TX 78712 USA
关键词
Single and multi-agent learning techniques; Reinforcement Learning; Factored Domains; Model Learning; Hierarchical Skill Learning; Monte Carlo Methods;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Reinforcement learning (RL) is a well-established paradigm for enabling autonomous agents to learn from experience. To enable RL to scale to any but the smallest domains, it is necessary to make use of abstraction and generalization of the state-action space, for example with a factored representation. However, to make effective use of such a representation, it is necessary to determine which state variables are relevant in which situations. In this work, we introduce T-UCT, a novel model-based RL approach for learning and exploiting the dynamics of structured hierarchical environments. When learning the dynamics while acting, a partial or inaccurate model may do more harm than good. T-UCT uses graph-based planning and Monte Carlo simulations to exploit models that may be incomplete or inaccurate, allowing it to both maximize cumulative rewards and ignore trajectories that are unlikely to succeed. T-UCT incorporates new experiences in the form of more accurate plans that span a greater area of the state space. T-UCT is fully implemented and compared empirically against B-VISA, the best known prior approach to the same problem. We show that T-UCT learns hierarchical models with fewer samples than B-VISA and that this effect is magnified at deeper levels of hierarchical complexity.
引用
收藏
页码:1985 / 1986
页数:2
相关论文
共 50 条
  • [1] Monte Carlo Hierarchical Model Learning
    Menashe, Jacob
    Stone, Peter
    PROCEEDINGS OF THE 2015 INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS & MULTIAGENT SYSTEMS (AAMAS'15), 2015, : 771 - 779
  • [2] Monte Carlo Bayesian Hierarchical Reinforcement Learning
    Ngo Anh Vien
    Hung Ngo
    Ertel, Wolfgang
    AAMAS'14: PROCEEDINGS OF THE 2014 INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS & MULTIAGENT SYSTEMS, 2014, : 1551 - 1552
  • [3] HIERARCHICAL MONTE-CARLO SIMULATION OF THE ISING-MODEL
    FAAS, M
    HILHORST, HJ
    PHYSICA A, 1986, 135 (2-3): : 571 - 590
  • [4] Monte Carlo Sampling with Hierarchical Move Sets: POSH Monte Carlo
    Nilmeier, Jerome
    Jacobson, Matthew P.
    JOURNAL OF CHEMICAL THEORY AND COMPUTATION, 2009, 5 (08) : 1968 - 1984
  • [5] Hierarchical Monte-Carlo Planning
    Ngo Anh Vien
    Toussaint, Marc
    PROCEEDINGS OF THE TWENTY-NINTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2015, : 3613 - 3619
  • [6] Hierarchical Monte Carlo image synthesis
    Keller, A
    MATHEMATICS AND COMPUTERS IN SIMULATION, 2001, 55 (1-3) : 79 - 92
  • [7] MONTE-CARLO VERIFICATION OF THE OPERATOR MODEL OF LEARNING
    KOMENDA, S
    MACAKOVA, J
    HRBEK, J
    MOHAPL, J
    ACTIVITAS NERVOSA SUPERIOR, 1985, 27 (04): : 279 - 281
  • [8] On the Hierarchical Bernoulli Mixture Model Using Bayesian Hamiltonian Monte Carlo
    Suryaningtyas, Wahyuni
    Iriawan, Nur
    Kuswanto, Heri
    Zain, Ismaini
    SYMMETRY-BASEL, 2021, 13 (12):
  • [9] Proposal of Consistent Learning Model with Exchange Monte Carlo
    Shibata, Hiroki
    Takama, Yasufumi
    2022 INTERNATIONAL CONFERENCE ON TECHNOLOGIES AND APPLICATIONS OF ARTIFICIAL INTELLIGENCE, TAAI, 2022, : 197 - 202
  • [10] Hierarchical Reinforcement Learning With Monte Carlo Tree Search in Computer Fighting Game
    Pinto, Ivan Pereira
    Coutinho, Luciano Reis
    IEEE TRANSACTIONS ON GAMES, 2019, 11 (03) : 290 - 295