Spatiotemporal Costmap Inference for MPC Via Deep Inverse Reinforcement Learning

被引:15
|
作者
Lee, Keuntaek [1 ]
Isele, David [2 ]
Theodorou, Evangelos A. [3 ]
Bae, Sangjae [2 ]
机构
[1] Georgia Inst Technol, Dept Elect & Comp Engn, Atlanta, GA 30318 USA
[2] Honda Res Inst USA Inc, Div Res, San Jose, CA 95110 USA
[3] Georgia Inst Technol, Sch Aerosp Engn, Atlanta, GA 30318 USA
关键词
Learning from demonstration; reinforcement learning; optimization and optimal control; motion and path planning; autonomous vehicle navigation;
D O I
10.1109/LRA.2022.3146635
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
It can he difficult to autonomously produce driver behavior so that it appears natural to other traffic participants. Through Inverse Reinforcement Learning (IRL), we can automate this process by learning the underlying reward function from human demonstrations. We propose a new IRL algorithm that learns a goal-conditioned spatio-temporal reward function. The resulting costmap is used by Model Predictive Controllers (MPCs) to perform a task without any hand-designing or hand-tuning of the cost function. We evaluate our proposed Goal-conditioned SpatioTemporal Zeroing Maximum Entropy Deep IRL (GSTZ)-MEDIRL framework together with MPC in the CARLA simulator for autonomous driving, lane keeping, and lane changing tasks in a challenging dense traffic highway scenario. Our proposed methods show higher success rates compared to other baseline methods including behavior cloning, state-of-the-art RL policies, and MPC with a learning-based behavior prediction model.
引用
收藏
页码:3194 / 3201
页数:8
相关论文
共 50 条
  • [21] Haptic Assistance via Inverse Reinforcement Learning
    Scobee, Dexter R. R.
    Royo, Vicenc Rubies
    Tomlin, Claire J.
    Sastry, S. Shankar
    2018 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2018, : 1510 - 1517
  • [22] Learning to Walk via Deep Reinforcement Learning
    Haarnoja, Tuomas
    Ha, Sehoon
    Zhou, Aurick
    Tan, Jie
    Tucker, George
    Levine, Sergey
    ROBOTICS: SCIENCE AND SYSTEMS XV, 2019,
  • [23] Bayesian Deep Reinforcement Learning via Deep Kernel Learning
    Junyu Xuan
    Jie Lu
    Zheng Yan
    Guangquan Zhang
    International Journal of Computational Intelligence Systems, 2018, 12 : 164 - 171
  • [24] Bayesian Deep Reinforcement Learning via Deep Kernel Learning
    Xuan, Junyu
    Lu, Jie
    Yan, Zheng
    Zhang, Guangquan
    INTERNATIONAL JOURNAL OF COMPUTATIONAL INTELLIGENCE SYSTEMS, 2019, 12 (01) : 164 - 171
  • [25] Autonomous Exploration of Mobile Robots via Deep Reinforcement Learning Based on Spatiotemporal Information on Graph
    Zhang, Zhiwen
    Shi, Chenghao
    Zhu, Pengming
    Zeng, Zhiwen
    Zhang, Hui
    APPLIED SCIENCES-BASEL, 2021, 11 (18):
  • [26] Advanced planning for autonomous vehicles using reinforcement learning and deep inverse reinforcement learning
    You, Changxi
    Lu, Jianbo
    Filev, Dimitar
    Tsiotras, Panagiotis
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2019, 114 : 1 - 18
  • [27] POMDP inference and robust solution via deep reinforcement learning: an application to railway optimal maintenance
    Arcieri, Giacomo
    Hoelzl, Cyprien
    Schwery, Oliver
    Straub, Daniel
    Papakonstantinou, Konstantinos G.
    Chatzi, Eleni
    MACHINE LEARNING, 2024, 113 (10) : 7967 - 7995
  • [28] Accuracy-Guaranteed Collaborative DNN Inference in Industrial IoT via Deep Reinforcement Learning
    Wu, Wen
    Yang, Peng
    Zhang, Weiting
    Zhou, Conghao
    Shen, Xuemin
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2021, 17 (07) : 4988 - 4998
  • [29] Deep reinforcement learning for inverse inorganic materials design
    Karpovich, Christopher
    Pan, Elton
    Olivetti, Elsa A.
    NPJ COMPUTATIONAL MATERIALS, 2024, 10 (01)
  • [30] Comfortable Driving by using Deep Inverse Reinforcement Learning
    Kishikawa, Daiko
    Arai, Sachiyo
    2019 IEEE INTERNATIONAL CONFERENCE ON AGENTS (ICA), 2019, : 38 - 43