Spatiotemporal Costmap Inference for MPC Via Deep Inverse Reinforcement Learning

被引:15
|
作者
Lee, Keuntaek [1 ]
Isele, David [2 ]
Theodorou, Evangelos A. [3 ]
Bae, Sangjae [2 ]
机构
[1] Georgia Inst Technol, Dept Elect & Comp Engn, Atlanta, GA 30318 USA
[2] Honda Res Inst USA Inc, Div Res, San Jose, CA 95110 USA
[3] Georgia Inst Technol, Sch Aerosp Engn, Atlanta, GA 30318 USA
关键词
Learning from demonstration; reinforcement learning; optimization and optimal control; motion and path planning; autonomous vehicle navigation;
D O I
10.1109/LRA.2022.3146635
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
It can he difficult to autonomously produce driver behavior so that it appears natural to other traffic participants. Through Inverse Reinforcement Learning (IRL), we can automate this process by learning the underlying reward function from human demonstrations. We propose a new IRL algorithm that learns a goal-conditioned spatio-temporal reward function. The resulting costmap is used by Model Predictive Controllers (MPCs) to perform a task without any hand-designing or hand-tuning of the cost function. We evaluate our proposed Goal-conditioned SpatioTemporal Zeroing Maximum Entropy Deep IRL (GSTZ)-MEDIRL framework together with MPC in the CARLA simulator for autonomous driving, lane keeping, and lane changing tasks in a challenging dense traffic highway scenario. Our proposed methods show higher success rates compared to other baseline methods including behavior cloning, state-of-the-art RL policies, and MPC with a learning-based behavior prediction model.
引用
收藏
页码:3194 / 3201
页数:8
相关论文
共 50 条
  • [1] Towards Interpretable Deep Reinforcement Learning Models via Inverse Reinforcement Learning
    Xie, Yuansheng
    Vosoughi, Soroush
    Hassanpour, Saeed
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 5067 - 5074
  • [2] Inverse Reinforcement Learning via Deep Gaussian Process
    Jin, Ming
    Damianou, Andreas
    Abbeel, Pieter
    Spanos, Costas
    CONFERENCE ON UNCERTAINTY IN ARTIFICIAL INTELLIGENCE (UAI2017), 2017,
  • [3] Active Task-Inference-Guided Deep Inverse Reinforcement Learning
    Memarian, Farzan
    Xu, Zhe
    Wu, Bo
    Wen, Min
    Topcu, Ufuk
    2020 59TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2020, : 1932 - 1938
  • [4] CBNAV: Costmap Based Approach to Deep Reinforcement Learning Mobile Robot Navigation
    Tomasi Junior, Darci Luiz
    Todt, Eduardo
    2021 LATIN AMERICAN ROBOTICS SYMPOSIUM / 2021 BRAZILIAN SYMPOSIUM ON ROBOTICS / 2021 WORKSHOP OF ROBOTICS IN EDUCATION (LARS-SBR-WRE 2021), 2021, : 324 - 329
  • [5] Costmap-based Local Motion Planning using Deep Reinforcement Learning
    Garrote, Luis
    Perdiz, Joao
    Nunes, Urbano J.
    2023 32ND IEEE INTERNATIONAL CONFERENCE ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION, RO-MAN, 2023, : 1089 - 1095
  • [6] Estimation of personal driving style via deep inverse reinforcement learning
    Kishikawa, Daiko
    Arai, Sachiyo
    ARTIFICIAL LIFE AND ROBOTICS, 2021, 26 (03) : 338 - 346
  • [7] Estimation of personal driving style via deep inverse reinforcement learning
    Daiko Kishikawa
    Sachiyo Arai
    Artificial Life and Robotics, 2021, 26 : 338 - 346
  • [8] Inverse reinforcement learning through logic constraint inference
    Mattijs Baert
    Sam Leroux
    Pieter Simoens
    Machine Learning, 2023, 112 : 2593 - 2618
  • [9] Inverse reinforcement learning through logic constraint inference
    Baert, Mattijs
    Leroux, Sam
    Simoens, Pieter
    MACHINE LEARNING, 2023, 112 (07) : 2593 - 2618
  • [10] Variational Inference MPC for Bayesian Model-based Reinforcement Learning
    Okada, Masashi
    Taniguchi, Tadahiro
    CONFERENCE ON ROBOT LEARNING, VOL 100, 2019, 100