Monte Carlo Hierarchical Model Learning

被引:0
|
作者
Menashe, Jacob [1 ]
Stone, Peter [1 ]
机构
[1] Univ Texas Austin, Austin, TX 78712 USA
关键词
Single and multi-agent learning techniques; Reinforcement Learning; Factored Domains; Model Learning; Hierarchical Skill Learning; Monte Carlo Methods;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Reinforcement learning (RL) is a well-established paradigm for enabling autonomous agents to learn from experience. To enable RL to scale to any but the smallest domains, it is necessary to make use of abstraction and generalization of the state-action space, for example with a factored representation. However, to make effective use of such a representation, it is necessary to determine which state variables are relevant in which situations. In this work, we introduce T-UCT, a novel model-based RL approach for learning and exploiting the dynamics of structured hierarchical environments. When learning the dynamics while acting, a partial or inaccurate model may do more harm than good. T-UCT uses graph-based planning and Monte Carlo simulations to exploit models that may be incomplete or inaccurate, allowing it to both maximize cumulative rewards and ignore trajectories that are unlikely to succeed. T-UCT incorporates new experiences in the form of more accurate plans that span a greater area of the state space. T-UCT is fully implemented and compared empirically against B-VISA, the best known prior approach to the same problem. We show that T-UCT learns hierarchical models with fewer samples than B-VISA and that this effect is magnified at deeper levels of hierarchical complexity.
引用
收藏
页码:1985 / 1986
页数:2
相关论文
共 50 条
  • [11] Hierarchical characterization of aggregates for Monte Carlo simulations
    Briesen, Heiko
    AICHE JOURNAL, 2006, 52 (07) : 2436 - 2446
  • [12] A survey of Monte Carlo algorithms for maximizing the likelihood of a two-stage hierarchical model
    Booth, James G.
    Hobert, James P.
    Jank, Wolfgang
    STATISTICAL MODELLING, 2001, 1 (04) : 333 - 349
  • [13] Sequential Monte Carlo Samplers for Model-Based Reinforcement Learning
    Sonmez, Orhan
    Cemgil, A. Taylan
    2013 21ST SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2013,
  • [14] An Enhanced Machine Learning Model for Adaptive Monte Carlo Yield Analysis
    Kimmel, Richard
    Li, Tong
    Winston, David
    PROCEEDINGS OF THE 2020 ACM/IEEE 2ND WORKSHOP ON MACHINE LEARNING FOR CAD (MLCAD '20), 2020, : 89 - 94
  • [15] Learning to Search Promising Regions by a Monte-Carlo Tree Model
    Xia, Hai
    Li, Changhe
    Zeng, Sanyou
    Tan, Qingshan
    Wang, Junchen
    Yang, Shengxiang
    2022 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC), 2022,
  • [16] MONTE CARLO STUDY OF ACCURACY OF A HIERARCHICAL GROUPING PROCEDURE
    GROSS, AL
    MULTIVARIATE BEHAVIORAL RESEARCH, 1972, 7 (03) : 379 - 389
  • [17] Monte Carlo shell model
    Otsuka, T
    NUCLEAR PHYSICS A, 2001, 693 (1-2) : 383 - 393
  • [18] Monte Carlo model checking
    Grosu, R
    Smolka, SA
    TOOLS AND ALGORITHMS FOR THE CONSTRUCTION AND ANALYSIS OF SYSTEMS, PROCEEDINGS, 2005, 3440 : 271 - 286
  • [19] On learning strategies for evolutionary Monte Carlo
    Goswami, Gopi
    Liu, Jun S.
    STATISTICS AND COMPUTING, 2007, 17 (01) : 23 - 38
  • [20] On learning strategies for evolutionary Monte Carlo
    Gopi Goswami
    Jun S. Liu
    Statistics and Computing, 2007, 17 : 23 - 38