Hierarchical Imitation Learning for Stochastic Environments

被引:0
|
作者
Igl, Maximilian [1 ]
Shah, Punit [1 ]
Mougin, Paul [1 ]
Srinivasan, Sirish [1 ]
Gupta, Tarun [1 ]
White, Brandyn [1 ]
Shiarlis, Kyriacos [1 ]
Whiteson, Shimon [1 ]
机构
[1] Waymo Res, Mountain View, CA 94043 USA
关键词
D O I
10.1109/IROS55552.2023.10341451
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Many applications of imitation learning require the agent to generate the full distribution of behaviour observed in the training data. For example, to evaluate the safety of autonomous vehicles in simulation, accurate and diverse behaviour models of other road users are paramount. Existing methods that improve this distributional realism typically rely on hierarchical policies. These condition the policy on types such as goals or personas that give rise to multi-modal behaviour. However, such methods are often inappropriate for stochastic environments where the agent must also react to external factors: because agent types are inferred from the observed future trajectory during training, these environments require that the contributions of internal and external factors to the agent behaviour are disentangled and only internal factors, i.e., those under the agent's control, are encoded in the type. Encoding future information about external factors leads to inappropriate agent reactions during testing, when the future is unknown and types must be drawn independently from the actual future. We formalize this challenge as distribution shift in the conditional distribution of agent types under environmental stochasticity. We propose Robust Type Conditioning (RTC), which eliminates this shift with adversarial training under randomly sampled types. Experiments on two domains, including the large-scale Waymo Open Motion Dataset, show improved distributional realism while maintaining or improving task performance compared to state-of-the-art baselines.
引用
收藏
页码:1697 / 1704
页数:8
相关论文
共 50 条
  • [31] HiLITE: Hierarchical and Lightweight Imitation Learning for Power Management of Embedded SoCs
    Sartor A.L.
    Krishnakumar A.
    Arda S.E.
    Ogras U.Y.
    Marculescu R.
    IEEE Computer Architecture Letters, 2020, 19 (01): : 63 - 67
  • [32] LEARNING BEHAVIORS OF THE HIERARCHICAL STRUCTURE STOCHASTIC AUTOMATA UNDER THE NONSTATIONARY MULTITEACHER ENVIRONMENTS AND THEIR APPLICATIONS TO INTELLIGENT ROBOT MANIPULATORS
    BABA, N
    IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS, 1987, 17 (05): : 868 - 873
  • [33] Virtual Institutions: Normative Environments Facilitating Imitation Learning in Virtual Agents
    Bogdanovych, Anton
    Simoff, Simeon
    Esteva, Marc
    INTELLIGENT VIRTUAL AGENTS, PROCEEDINGS, 2008, 5208 : 456 - +
  • [34] Active Learning within Constrained Environments through Imitation of an Expert Questioner
    Bullard, Kalesha
    Schroecker, Yannick
    Chernova, Sonia
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 2045 - 2052
  • [35] Reinforcement imitation learning for reliable and efficient autonomous navigation in complex environments
    Kumar D.
    Neural Computing and Applications, 2024, 36 (20) : 11945 - 11961
  • [36] Imitation Learning for Natural Language Direction Following through Unknown Environments
    Duvallet, Felix
    Kollar, Thomas
    Stentz, Anthony
    2013 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2013, : 1047 - 1053
  • [37] Pathfinding in stochastic environments: learning vs planning
    Skrynnik A.
    Andreychuk A.
    Yakovlev K.
    Panov A.
    PeerJ Computer Science, 2022, 8
  • [38] Pathfinding in stochastic environments: learning vs planning
    Skrynnik, Alexey
    Andreychuk, Anton
    Yakovlev, Konstantin
    Panov, Aleksandr
    PEERJ COMPUTER SCIENCE, 2022, 8
  • [39] Hierarchical Learning Approach for One-shot Action Imitation in Humanoid Robots
    Wu, Yan
    Demiris, Yiannis
    11TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION, ROBOTICS AND VISION (ICARCV 2010), 2010, : 453 - 458
  • [40] Imitation Learning of Hierarchical Driving Model: From Continuous Intention to Continuous Trajectory
    Wang, Yunkai
    Zhang, Dongkun
    Wang, Jingke
    Chen, Zexi
    Li, Yuehua
    Wang, Yue
    Xiong, Rong
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (02): : 2477 - 2484